Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "new cpu"
-
Hey everyone,
Merry Christmas to everyone who celebrates, happy holidays to everyone, and happy almost-new-year!
Tim and I wanted to reflect on the year devRant has had, and looking back, there are a lot of awesome things that happened in 2018 that we are very thankful for. Here are just a few of the ones that we thought of (this list is not exhaustive and I'm definitley forgetting stuff, so please comment about those!):
- After nearly a year in the making, the completely overhauled devRant web version was launched (https://devrant.com/rants/1255714/...)
- @linuxxx became the first devRant user to hit 100,000++! (https://devrant.com/rants/1157415/...)
- We once again pulled off the greatest April fools joke everrrr (https://devrant.com/rants/1311206/...)
- @trogus started making awesome devComics and http://devcomics.com was launched
- We added a feature to allow rant filtering by post type (https://devrant.com/rants/1354275/...)
- We made it so avatars could have expressions! (https://devrant.com/rants/1563683/...)
- We had a booth at TechDay New York and got to meet some devRant users! (https://devrant.com/rants/1394067/...)
- We made major backend architectural improvements - including spinning up a special high-powered-CPU web server to handle avatar creation and make the creation process much faster (https://devrant.com/rants/1370938/...)
- App stability: mainly Android - we fixed crashes, did a push-notif overhaul, and tried to continue making the apps better and more stable
- A record amount of devRant meetups were held, and we couldn't be more proud about that, and we thank every person who organized one! (just a few: https://devrant.com/rants/1588218/... https://devrant.com/rants/1884724/... https://devrant.com/rants/1683365/... https://devrant.com/rants/1922950/...)
We had a busy year, and despite some things going on for us personally and some setbacks around those, we think this was a very productve year for devRant and that we are going in the right direction. We're continuing to constantly evaluate feedback from members of the community to decide where to take the app next. We're fully committed to improving the devRant community in 2019 and we have a lot of ideas about how we can do that. We're working on some things, but we're not really announcing them yet, so please sit tight for those :) In the meantime, feel free to let us know what you'd like to see improved/added the most as we always like to get updated feedback from the community.
As always, thank you everyone, and thanks for your amazing contributions to the devRant community!
Looking forward to 2019,
- David and Tim26 -
So I visited my aunt's house a few days back.
They heard I write code (read: Google and copy-paste from Stack Overflow), and wanted me to help repair their computers.
Saw my cousin playing Battlefield 4 at sub-20 fps with a GTX 750 Ti on the lowest settings. His fucking CPU fan was bust, and judging by the amount of dust in his system, I literally thought he was cramming dust in there on purpose.
After a heavy dusting, another stick of RAM (4GB -> 12GB), a new heatsink (CoolerMaster T400i), and a fresh copy of Windows 10 (along with Office, etc.), he could play games at 60+ fps again.
What do I get? Not even a fucking thank you. Just a "you done yet? I want to play video games."
I mean... Gee. Your cousin flew all the way to a new continent, spent his precious vacation time helping you out, and all he gets is a cold-ass shoulder.
Even my fucking ex gave me more than that.16 -
one of our computers at work suddenly shut down. our boss panicked like it was the end of the world cause he knew we couldnt buy a new one and we desperately need the computer. when i came in he started telling me he's gonna pay me extra if i fix the computer
me: *checks cpu and finds out it was unplugged* u sure bout that?
him: hell yea, id rather pay u that buy a new one
*30 mins later*
me: *plugs it back in and pushes power button* its done
>> guess who's got extra money without any extra efforts yay11 -
I'm getting ridiculously pissed off at Intel's Management Engine (etc.), yet again. I'm learning new terrifying things it does, and about more exploits. Anything this nefarious and overreaching and untouchable is evil by its very nature.
(tl;dr at the bottom.)
I also learned that -- as I suspected -- AMD has their own version of the bloody thing. Apparently theirs is a bit less scary than Intel's since you can ostensibly disable it, but i don't believe that because spy agencies exist and people are power-hungry and corrupt as hell when they get it.
For those who don't know what the IME is, it's hardware godmode. It's a black box running obfuscated code on a coprocessor that's built into Intel cpus (all Intell cpus from 2008 on). It runs code continuously, even when the system is in S3 mode or powered off. As long as the psu is supplying current, it's running. It has its own mac and IP address, transmits out-of-band (so the OS can't see its traffic), some chips can even communicate via 3g, and it can accept remote commands, too. It has complete and unfettered access to everything, completely invisible to the OS. It can turn your computer on or off, use all hardware, access and change all data in ram and storage, etc. And all of this is completely transparent: when the IME interrupts, the cpu stores its state, pauses, runs the SMM (system management mode) code, restores the state, and resumes normal operation. Its memory always returns 0xff when read by the os, and all writes fail. So everything about it is completely hidden from the OS, though the OS can trigger the IME/SMM to run various functions through interrupts, too. But this system is also required for the CPU to even function, so killing it bricks your CPU. Which, ofc, you can do via exploits. Or install ring-2 keyloggers. or do fucking anything else you want to.
tl;dr IME is a hardware godmode, and if someone compromises this (and there have been many exploits), their code runs at ring-2 permissions (above kernel (0), above hypervisor (-1)). They can do anything and everything on/to your system, completely invisibly, and can even install persistent malware that lives inside your bloody cpu. And guess who has keys for this? Go on, guess. you're probably right. Are they completely trustworthy? No? You're probably right again.
There is absolutely no reason for this sort of thing to exist, and its existence can only makes things worse. It enables spying of literally all kinds, it enables cpu-resident malware, bricking your physical cpu, reading/modifying anything anywhere, taking control of your hardware, etc. Literal godmode. and some of it cannot be patched, meaning more than a few exploits require replacing your cpu to protect against.
And why does this exist?
Ostensibly to allow sysadmins to remote-manage fleets of computers, which it does. But it allows fucking everything else, too. and keys to it exist. and people are absolutely not trustworthy. especially those in power -- who are most likely to have access to said keys.
The only reason this exists is because fucking power-hungry doucherockets exist.26 -
Friend: I bought a new phone.
Me: nice!
Friend: It's really fast coz it has 3gb ram!!
Me: okay... Other specs? Cpu? Camera?
Friend: 13 MEGAPIXELS!
Me: yeah but other specs?
Friend: doesn't matter! The photos are of 8mb!
Me: yeah. Best phone ever. 😑😓10 -
Sooo, in my 5 years of high school, I had 5 different IT teachers...
Now, in Italy Highschool goes from 14 to 19 years old, I started programming some days after becoming 13, and "programming" classes begin on the third year, so I had quite a headstart on my classmates...
Now, for the third year, I had an awesome teacher, he noticed I was ahead and... Bored, so he gave me some extra stuff to study, he's the only teacher I've learnt anything from, it was awesome, very stingy with grades, but getting a perfect score with him was so satisfying.
Fourth year, the new guy was old, very old, at least 70, his lessons were just him talking about how programming was when he was young.
But then... During the second half of the fourth year I changed class due to bullying under a teacher's advice, and HE happened...
My new IT teacher, one of the most ignorant, awful people I ever met...
He's literally the reason I only went back to that school once, because another teacher needed help with a course...
One day I made the HUGE mistake to say that his "while(i <10000000000000);" wasn't very efficient for making a delay, because it didn't free the CPU, and since then:
- I never got more than 7 out of 10 at his tests
- He insulted me in front of the whole class
- He sabotaged the oral part of my final exam, shouting that he hated D'Annunzio when he saw he was in the literature part of my thesis (needed him to connect to WW2, and the Memex, that then allowed me to start talking about PCs and programming, my thesis was about the influence of lisp on modern programming languages), loudly chatting with other teachers when I was trying to keep calm (a teacher who knows me quite well, and was there to see my "performance" thought I was going to snap at some point), distracting the english teacher when I was exposing the english part of my thesis and pressuring the commission to give me 99 instead of 100 out of 100
So yeah, he almost made me hate the only thing I'm good at, undervaluing my work and my skills, undervaluing and humiliating me as a person, and I think that if I meet him again I might spit on his face...
So yeah, my biggest "programmer enemy" was a person that then did everything in his power to make my last year and a half of highschool hell
Now I can gladly say that with the help of my tutoring, some of my university colleagues are starting to appreciate programming, and my engineer friends ask for my help when they need advices about their code, and it's giving me motivation to keep doing it and becoming a better programmer to keep up with their expectations4 -
"For our app we are using Docker, so for the new server we will need a barebone Linux install"
"Oh please no, Linux is not for a stable server and is very hard to manage, let's install Windows Server 2002 and you can virtualize everything you want"
"Wait, we really can use our cpu cycles to something better..."
"Sorry, too late, already installing!"
Our sysadmin is a Microsoft integralist, we should RUN!18 -
Paypal Rant #2
Paypal might just be the only company with 98% of their employees being support staff because not a soul on this planet knows how to work with that fucking piece of shit of a company's service.
No really, if there was a shittiness-rating from 1-10 (10 = worst) you would have to store paypal's rating as a string or invent a new data type because no CPU could fucking work with such a big ass fucking number.
If I had to choose between Paypal and going back to manually trading physical goods/animals for stuff I would gladly choose the latter, because Paypal, go suck a bag of dicks you useless fucking shitpile of a "company".8 -
Year 2200, somewhere unknown. A robot drone flies above a now destroyed city. As it slowly descends down, it finds it's way into a building. In the rubble, a monitor is still flashing. The drone moves toward the monitor. After few minutes of analysis, drone releases a connector from it's belly. It connects to CPU nearby. After scanning the system for another few minutes, it's apparent the system was not the new OS which the Government released and Mandated. Rather, a prehistoric OS which legends spoke in hushed tones. An OS so powerful, which controlled then millions of computers.
Like the story so far? Should I continue?9 -
so my mom wanted to write some word document, but she didn't use her laptop for like ~5 years, it didn't boot up so she asked me to fix, now here is what I found :
>the laptop had a 240 gb hdd
>the hdd was literally broken
>bought her a new 500 gb hdd
>installed windows 7
>took 10 mins to install
>took 19 minutes to boot up
>removed windows 7
>installed win xp
>took 30 mins to install
>took 3 minutes to boot up
>opened windows
>checked pc specs
>see picture below
>[insert wtf gif here]
>installed drivers
>took 20 minutes to install drivers
>[insert epic music here]
>tried installing office 2016
>insta regret
>tried installing office 2010
>memory farted and I couldn't even move the cursor
>installed office 2007
>mom started writing document peacefully
>after 2 hours bsod
>mom asks me to fix
>opens laptop to check internal components
>the cpu had a black hole inside
>the fans weren't working due to the circuit being burnt for some reason
>kills laptop
>kills mom
>kills self
>live peacefully in hell11 -
Hello again, everyone. As Sunday comes to a close, and Monday is fast approaching, I'll share with you the likely cause of my death by stroke and/or heart attack:
MONDAY MORNING COFFEE OF HORROR
Disclaimer: Do NOT try this. I am a professional addict. I am not responsible for anything this brew from hell causes to you and/or those around you.
So, I wake up, feeling like I haven't slept for days, or just notice the fucking alarm clock shrieking because I pulled an all-nighter.
Step 1: Silence alarm clock via mild violence.
Step 2: Get the coffee machine to brew some filter coffee (espresso works too)
Step 3: Get milk and ice cubes from the fridge (both are needed, I don't care if you don't like milk, trust me)
Step 4: Get 2 spoonfuls (not tea spoon, and actually FULL spoonfuls) into the biggest glass you have
Step 5: Pour just a little of the warm filter coffee into the glass, just to get the instant coffee wet enough, and start mixing, until the result looks like the horror you unleashed in your toilet a few minutes ago (and will do so again in a few)
Step 6: Mix in 25-50 ml milk, just for the aesthetic change of colour of the devil-brew, and to add the necessary amount of lactic acid to react with the coffee to produce chemical X
Step 7: Add ice cubes to taste (if you are new to this, add a lot)
Step 8. Slowly add the filter coffee while mixing furiously, so that the light brown paste at the bottom get dissolved (it's harder than it sounds)
Now, take a deep breath. Before you is a disgusting brew undergoing a chemical reaction, and your moves need to be precise otherwise it will explode. Note that sugar or any other form of sweetener is FORBIDDEN, as it will block the reaction chain and the result won't be as potent.
Take a straw (a big one, not those needle-like ones that some cafeterias give to fool you into believing that the coffee is more than 150ml). Put it inside the mix, and check that the route to the bathroom is free of obstacles.
Now, clench your abs, close your nose if you are new to this, grab the straw and DRINK!
DRINK LIKE THERE IS NO TOMORROW!
THAT BROWN DEVIL'S BILE WILL HAVE YOUR INTESTINES SPASM AND DANCE THE MACARENA WHILE TWIRLING A HULA HOOP!
YOUR HEART WILL GO OVERDRIVE HARDER THAN YOUR PC'S CPU WHEN COMPILING ON ECLIPSE AND BROWSING WITH IE AT THE SAME TIME.
The combination of caffeine and lactic acid will bring out the perfectly disgusting combination of sour and bitter usually expected in rotting lemons. After you manage to chug it down (DON'T SPILL OR SPIT ANY!) you have 30 - 60 seconds max to run to the porcelain throne, where you will spend the next 30-60 minutes.
After that, nothing can stop you! You will fix bugs, write entire codebases from scratch, punch that annoying coworker, punch that boss! You will be a demigod among mortals for the next 6-8 hours!
Your recipes for Monday morning coffee?15 -
Most satisfying bug I've fixed?
Fixed a n+1 issue with a web service retrieving price information. I initially wrote the service, but it was taken over by a couple of 'world class' monday-morning-quarterbacks.
The "Worst code I've ever seen" ... "I can't believe this crap compiles" types that never met anyone else's code that was any good.
After a few months (yes months) and heavy refactoring, the service still returned price information for a product. Pass the service a list of product numbers, service returns the price, availability, etc, that was it.
After a very proud and boisterous deployment, over the next couple of days the service seemed to get slower and slower. DBAs started to complain that the service was causing unusually high wait times, locks, and CPU spikes causing problems for other applications. The usual finger pointing began which ended up with "If PaperTrail had written the service 'correctly' the first time, we wouldn't be in this mess."
Only mattered that I initially wrote the service and no one seemed to care about the two geniuses that took months changing the code.
The dev manager was able to justify a complete re-write of the service using 'proper development methodologies' including budgeting devs, DBAs, server resources, etc..etc. with a projected year+ completion date.
My 'BS Meter' goes off, so I open up the code, maybe 5 minutes...tada...found it. The corresponding stored procedure accepts a list of product numbers and a price type (1=Retail, 2=Dealer, and so on). If you pass 0, the stored procedure returns all the prices.
Code basically looked like this..
public List<Prices> GetPrices(List<Product> products, int priceTypeId)
{
foreach (var item in products)
{
List<int> productIdsParameter = new List<int>();
productIdsParameter.Add(item.ProductID);
List<Price> prices = dataProvider.GetPrices(productIdsParameter, 0);
foreach (var price in prices)
{
if (price.PriceTypeID == priceTypeId)
{
prices = dataProvider.GetPrices(productIdsParameter, price.PriceTypeID);
return prices;
}
* Omitting the other 'WTF?' code to handle the zero price type
}
}
}
I removed the double stored procedure call, updated the method signature to only accept the list of product numbers (which it was before the 'major refactor'), deployed the service to dev (the issue was reproducible in our dev environment) and had the DBA monitor.
The two devs and the manager are grumbling and mocking the changes (they never looked, they assumed I wrote some threading monstrosity) then the DBA walks up..
DBA: "We're good. You hit the database pretty hard and the CPU never moved. Execution plans, locks, all good to go."
<dba starts to walk away>
DevMgr: "No fucking way! Putting that code in a thread wouldn't have fix it"
Me: "Um, I didn't use threads"
Dev1: "You had to. There was no way you made that code run faster without threads"
Dev2: "It runs fine in dev, but there is no way that level of threading will work in production with thousands of requests. I've got unit tests that prove our design is perfect."
Me: "I looked at what the code was doing and removed what it shouldn't be doing. That's it."
DBA: "If the database is happy with the changes, I'm happy. Good job. Get that service deployed tomorrow and lets move on"
Me: "You'll remove the recommendation for a complete re-write of the service?"
DevMgr: "Hell no! The re-write moves forward. This, whatever you did, changes nothing."
DBA: "Hell yes it does!! I've got too much on my plate already to play babysitter with you assholes. I'm done and no one on my team will waste any more time on this. Am I clear?"
Seeing the dev manager face turn red and the other two devs look completely dumbfounded was the most satisfying bug I've fixed.5 -
Would the web be better off, if there was zero frontend scripting? There would be HTML5 video/audio, but zero client side JS.
Browsers wouldn't understand script tags, they wouldn't have javascript engines, and they wouldn't have to worry about new standards and deprecations.
Browsers would be MUCH more secure, and use way less memory and CPU resources.
What would we really be missing?
If you build less bloated pages, you would not really need ajax calls, page reloads would be cheap. Animated menus do not add anything functionally, and could be done using css as well. Complicated webapps... well maybe those should just be desktop/mobile apps.
Pages would contain less annoying elements, no tracking or crypto mining scripts, no mouse tracking, no exploitative spam alerts.
Why don't we just deprecate JS in the browser, completely?
I think it would be worth it.22 -
Just added Kotlin plugin to my new android project and what do you know, CPU went to 100% of usage.
Aaand the winner is: Fucking Android Studio!19 -
I did it: I built up another PC identical to my machine (https://devrant.com/rants/2923002/...) for my SO and installed Linux Mint for her, too. That had been my primary motive for an easy and stable distro in the first place.
Now that didn't come out of the blue. We were discussing the end of Win 7 already two years ago where I brought up my concerns with Win 10 - mainly the forced, lousy updates and the integrated spyware, and that I was considering Linux as way out.
I had expected quite some pushback because she had been exclusively on Windows since the 90s. However, I didn't sell Linux as upgrade. It's just that Win 7 is over, progress under Windows as well, and we're in damage control mode. Went down pretty well.
Fast forward three weeks - remember, first time Linux user and no IT-geek:
- it just works, including web, videos, and music.
- she likes Cinnamon.
- nice desktop themes.
- Redshift is as good as f.lux.
- software installation is just like an app store.
- updates work via an easy tray icon.
- quote: "Linux is great!"
- given this alternative, she doesn't understand why people willingly put up with Win 10.
- no drive letters: already forgotten.
- popcorn for upcoming Win 10 disaster stories.
- why do Windows updates take that long?
- why does Windows need to reboot for every update?
- why does Windows hang in that update boot screen for so long?
I'm impressed that Linux has come so far that it's suitable for end users. Next in line is her father who wants to try Linux, but that will be a story for tomorrow.22 -
My "new desktop PC".
6 core xeon CPU with hyperthreading x2 (24 virtual cores) over clocked to 3.9ghz
128gb ddr3 ecc ram
256gb SCSI hard drives in RAID 0
256 Samsung ssd
AMD r9 something something 8gb ddr5 vram gpu
Got it for £120 on eBay (except for the ssd and gpu). Still missing a sound card :(
I have no idea on what to use it yet except for development but I thought it'd be cool to build a pretty good "desktop pc" for around £350-£400.14 -
I realized this week that my CPU is almost 12 years old!
It's an AMD FX-8150 8core @3.6 (recently overclocked to 4ghz). It's still a good chip for my needs, but eep. I had no idea I bought it that long ago.
I really want to replace it, but that would require a new mobo, too. which I suppose wouldn't be a bad thing...
------
Unpopular onion:
AMD was comparable to (and occasionally better than) Intel before, but now? Ever since they bought ATI, they've just spiraled: virtually everything they've produced has lagged behind.
However: the Ryzen seems to break this trend, so maybe there's hope yet?undefined ryzen anachronism much? cpu intel seriously though: gtx970ssc + amd8150? 8150 unpopular onion he's a dorky one outmoded tech amd10 -
Intel: our new 12th gen CPU beats Ryzen in gaming!
Also Intel: benchmarks on a buggy version of Windows 11 that takes a big performance hit on Ryzen.
Yeah, Intel... you and your benchmarks. It's just so ridiculous.10 -
Here's a genuine rant for you. Probably the only one I've ever made and ever will..
It's a bit depressing and covers a few topics so just read it, it's important.
*deep inhale*
So, with the help of my friend and my Nana, I was getting VR set up. (Oh, what joy.)
Now, I love everything about VR. But the thing is, I've had this damned headset since may (Dell MMR) and I haven't been able to use it. The reason for that is, something always came up that I needed to buy and this became a huge deal.
But let's start from the beginning.. I'm curentally fighting depression. I have been for months. My only income is what my Nana gives me ($150/mo) and what my friend ocasinally gives me.
Anyways, the first issue was that I couldn't afford the headset. This was find, as my friend would get it for me, and I would pay them back the following month. But, then, once I got the headset that's when the real problems started. First it was that I needed bluetooth, so I bought an adapter. Then I realized my entire CPU was incompatable, so I had to get a new tower and I went ahead and got a new GPU as well. I also got a charging kit for my headset (This ended up making me owe my Nana money). Then after all of that was settled, I learned that the evauation software lied, and my computer doesn't have USB 3, so I need that too but low and behold; both of my graphics cards cover my second pcie slot. So my options are to either try and rig up something, or to buy a cpu and psu for my third AMD PC which I had forgotten about during this whole ordeal..
This was soposed to help me with my depression and stress. Now I don't even want to get out of bed.
With all that said, I might be getting on SSI soon (I'm sure some of you are familar with that, and no I don't want to talk about it) and when that happens I might just leave behind tech (well, my PC and games) and all the stress and pain it's caused me over my life so this was all for nothing.
Honestly.. I'm just done with everything. To all the new faces around here; Hello! How are ya? To everyone else; You know me. I've been around for a while, though I'm not popular because I lurked and commented with Alice. You all probably noticed that I left a while back, and it was because I was trying to get out of tech. My reason for tech was that I was searching for something. I was always looking for the next game to sate me, or fill this gap in my life. I became a programmer because it gave me control were I lacked it otherwise. I made friends online because my anxiety prevented me from doing so in the real world.
But to what end? What have I acomplished? My twenty second birthday is next month. I've no job, I move from family member to famly member because I'm so fixated on becoming someone else to make something of myself.
I have my own ideals, but it seems that I push them aside to try (and fail to) impress others.
It's time for change. Of course, I can't do anything without money, so I'll have to wait for my SSI which I will get news on in August.
I hope this message came through how I meant it to. There is so much I want to say, but I've no words to say them. And btw, the VR thing is just one of manny issue that i've delt with (but certanly the most expensive)
Alice, Zennoe (Alexis, whom is not on devrant); I'm not giving up tech entirely. don't expect to suddenly not hear from me. I'm mostly just giving up my computer and games. More casually so for now, and them more seriously once I get on SSI. I'll still message you every/other day like I have been. <326 -
*wondered for 4 years how a bootloop looks like*
Nexus: yOU wAnE kNoW wHaT a BoOtLoOp LoOkS LiKe?!
*bootloops itself to shit*
Well I guess that I know what I'll be doing tonight then. Flash that new StatixOS build because the phone shat itself.
*tries to reflash the recovery*
*still bootloops*
*tries to flash the stock OS*
*still fucking bootloops*
*finds a post on XDA saying something about fucked big cores that need to be disabled*
Fucking piece of junk. So not only the battery is shit, but also the CPU is shit, huh. Certified pieces of shit.
*flashes the patched boot.img that disables the big cores*
*phone loads Google logo.. good*
*BOOTLOOPS FUCKING AGAIN*
MJHUIETHNIUBESZPTUIBG ESVGU d
FUCK!!! Fuck you Google, fuck you Nexus, fuck you Huawei, HOW DIFFICULT CAN IT BE TO DESIGN A FUCKING PHONE?!!!
So yeah. Looking for suggestions for a new phone. Anything of which the kernel source is released and of which the battery is halfway decent (unlike this fucking piece of shit) should do.7 -
Have been trying to setup Netdata as a monitoring system for a while now and finally got it working!
Instead of the built-in webhooks I just did a curl to a url containing a php page/file which error logs the status and description (just for testing).
It took me way too long to get it to work but BAM.
Immediately made a new cpu load rule (one minute high load):
The satisfaction of getting an error message in the php logs containing my custom rule as warning and a minute later as critical 😍
Netdata ❤6 -
I'm 4 days into my new job, and so far I am absolutely loving it. Here's my setup. Yes, they gave me 3 monitors plus a laptop, so my setup has 5 screens! Now I can die happy :D
Definitely worth noting as well, since it caught me by surprise - the company-supplied laptop is a powerhouse. High-end i7 CPU, mid-to-high-end NVIDIA GPU, tons of ports, 1TB HD, 4K display, and 48 GB RAM. Yes, 48 GB. I am truly blessed, starting off my career with this. ^_^rant excessive ram my dream finally came true hello innumerable open tabs multi-monitor setup goodbye lag11 -
I don't know which should make me worry:
1. the fact a gradle build for a new project is taking 90% of CPU
2. The fact CPU Freq isn't going higher than 2.18 when it can go up to 3.1?
3. The fact gradle is always taking 5 minutes to build when it used to take 1 minute for a medium project?15 -
So today I got a new CPU after the old ones were decommissioned. A fresh new desktop.
"Where is my chrome I think."
"Oh ya", I open Internet Explorer, type Google.
Google delightfully asks," Want to try chrome?."
"Yes sir I do."
-This one never gets old!!15 -
Step 1: Run to the store to buy a USB card reader because all of a sudden you have a need to use a 16Mb CF card that was tossed in a junk drawer for 20 years (hoping it still works, of course), but that was the easy part...
Step 2: Realize that the apps - your own - you want to run on your new (old) Casio E-125 PocketPC (to re-live "glory" days) are compiled in ARM format, not MIPS, which is the CPU this device uses, and the installer packages you have FOR YOUR OWN APPS don't include MIPS, only ARM (WHY DID I DO THAT?!), so, the saga REALLY begins...
Step 3: Get a 20-year old OS to install in a Hyper-V VM... find out that basic things like networking don't work by default because the OS is so damn old, so spend hours solving that and other issues to get it to basically run well enough to...
Step 4: Get that OS updated so that it's at least kind/sorta/maybe (but between you and me, not really!) safe online, all without a browser that will work on ANY modern site (oh, and good luck finding a version of Firefox that runs on it - that all took a few hours)...
Step 5: Okay, OS is ready to go, now get 20-year old dev tools that you haven't even seen in that many years working. Oh, do this with a missing CD key and ISO's that weren't archived in a format that's usable today, plus a bunch of missing dependencies because the OS is, again, SO old (a few MORE hours)...
Step 6: Get 20-year old code written in a language you haven't used in probably almost that long to compile, dealing with pathing issues, missing libs, and several other issues, all the while trying to dust off long-dormant knowledge somewhere in the deep, dark recesses of your brain... surprisingly, it all came back to me, more or less, in under an hour, which lead to...
Step 7: FINALLY get it all to work, FINALLY get the code to compile, FINALLY get it transferred to the device (which has no network capabilities, by the way, which is where the card reader and CF card came into play) and re-live the glory of your old, crappy PocketPC apps and games running on the real thing! WOO-HOO!
Step 8: Realize it's 3:30am by the time that's all done and be VERY thankful that you're on vacation this week or work tomorrow would SSUUCCKK!!!!
Step 9. Get called into work the next day for a production issue despite being tired from the night before and an afternoon of errands, lose basically a whole day of vacation (7 hours spent on it) and not actually resolve it by after midnight when you finally say that's enough :(
Talk about your highs and your lows.6 -
Imac - brand new <16hrs used: Disk error, probably needs to return to apple
iPhone <6 months old: loses cover
iPhone case 2hours of use: loses color
Macbook Pro 2017: randomly shuts down...
Mac mini: CPU halted error
Other iMac (old one): Thread error...
Apple, what the f is going on??? Software problems with root etc. But now, what the f!!8 -
PORTFOLIO INFLATION
when every junior is writing algorithms, the next step up, the only way to keep up is writing apps. When every junior is writing apps, the next leg up is writing an entire SN.
Eventually junior full stack devs are writing microservice streaming cloud backend content delivery optimized social networks wrapped in virtualization with load balancing, proper CI, public accessible analytics apis, written in custom webaseembly compiled scripting backend utilizing both the latest graphql and every single feature of postgres, while also being a web site builder, an in browser app, mobile optimized, designed to transmogrify your asset pipelines linearflow functional-oriented modular rust cratified turbencabulator while cooking your turducken with CPU cycles, diffusing your gpt, and finetunning your llama 69 trillion parameter AI model to jerk you off all at the same time.
And then the title "wizard" becomes a reality as the void of meaning in our lives occupied by the anxiety of trying to reduce the fear of rejection in job hunting, is subsumed by the brief accidental glance into the cthulian madness-inducing yawning abyss of the future which is all the rest of our lives we have to endure existing for until at last sweet sweet death consumes us and we go to annihilation never having to configure one more framework or devops deploy of another virtual environment.
And it dawns on us that we no longer develop or write code at all. No, everything has become a "service" in this new hellscape future. We slowly come to the realization that every job is really just Costco greeter, or eventually going to be reduced to something equivalent, all human creativity, free will and emotions now taken care of by the automation while we manage the human aspects, like sardines pushing against one another not realizing their doom has been sealed along with the airless can they have been packed into, to be suffocated by circumstance and a system designed to reduce everything to a competition of metrics designed by the devil, if the metrics were misery", and "torture", while we ourselves are driven by this ratfuck wheel to turn endlessly toward social cannibalism, like rats eating their babies, but for the amusement of wallstreet corporate welfare whores who couldnt turn a dime if it wasnt already stolen.
And on our gravestones, those immortal words are carved, by the last person who gave up the ghost, the last whose soul wasnt yey shovelled onto the coal fires driving the content machine consuming the world:
Welcome to costco. I love you.12 -
Forgot to post a book yesterday, so maybe I’ll post two books today...
Anyway, this book, I found it recently never seen it before. But boy is it great.
It’s similar to the programming pearls book as far as what it’s about. Think of the refactoring book, clean code, programming pearls, and the mythical man month books, thrown in a blender, added some new spice and some new things, and filtered down into 100 or so page book, simple quick and enjoyable actually.
This book the references staple books by Sedgwick, knuth, Brooks, Myers, and so many others. It’s funny how things come full circle.
My favorite quote from the book. I’ve been essentially saying this for years, but to see it on a book, it’s lovely... more people need to realize it too.
“Understanding how things work at a low level becomes a base for making good decisions at the high level”
Followed up with if you’ve never built a computer from scratch your missing out... get yourself a breadboard and some TTL logic.. and build a 4 bit CPU, once you know how to program in assembly the next step is building your own computer ... if your university didn’t teach a class that did this they ripped you off....
Don’t bitch at me.. the book said that.. and I agree! 100% because it’s true, you can’t debate that.
Oh and btw this is another book written by a female developer.. kudos to her for nailing so many topics in such a short book!35 -
WHY THE FUCK! WHY THE FUCKING FUCK! DO I HAVE TO WAIT 3 FUCKING DAYS TO GET A FUCKING VIDEO RENDERED! i didnt buy a new fucking 2080TI for this! WHY THE FUCK DOES THE CPU RENDER THE FUCKING VIDEO!
I mean, we can do fucking REAL TIME RAY-TRACING! And yet, no fucking idiot came to the Idea, "hmm we could let the GPU its intended purpose and dont use the CPU that much." I MEAN, IT HAPPENED, BUT FUCK IT! FUCK ALL OF THIS! FUCKING 74 HOURS!! FOR AN HOUR CLIP!
(Its 4K tho)
Fuck.21 -
Haha kids, you're all dead wrong. Here's my story.
There is a thing called “emergence”. This is a fundamental property of our universe. It works 100% of the time. It can't be stopped, it can't be mitigated. Everything you see around you is an emergent phenomenon.
Emergence is triggered when a lot of similar things come together and interact. One water molecule cannot be dry or wet, but if you have many, after a certain number the new property emerges — wetness. The system becomes _wet_.
Professionalism is an emergent phenomenon too, and its water molecules are abstract knowledge. Learn tech things you're interested in, complete random tutorials, code, and after a certain amount of knowledge molecules is gained, something clicks inside your head, and you become a professional.
Unfortunately, there are no shortcuts here. Uni education can make you a professional seemingly quicker, but it's not because uni knowledge is special, it's because uni is a perfect environment to absorb a lot of knowledge in a short period of time.
It happened to me too. I started coding in Pascal in fifth grade of high school, and I did it till sixth. Then, seventh to ninth were spent on my uni's after-school program. After ninth grade, I drop out of high school to get to this uni's experimental program. First grade of uni, and we're making a CPU. Second grade, and we're doing hard math, C and assembly.
And finally, in the third grade, it happens. I was sitting there in the classroom, it was late, and I was writing a recursive sudoku solver in Python. And I _felt_ the click. You cannot mistake it for anything else. It clicks, and you're a changed person. Immediately, I realized I can write everything. Needless to say, I was passing everything related to code afterwards with flying colours.
From that point, everything I did was just gaining more and more experience. Nothing changed fundamentally.
Emergence is forever. If you learn constantly, even without a concrete defined path, I can guarantee you that you _will_ become a professional. This is backed by the universe itself. You cannot avoid becoming one if you're actively accumulating emergence points.
Here's the list of projects I made in the past 11 years: https://notion.so/uyouthe/...
I'm 24.7 -
There's just way too many choices when it comes to buying a new pc. And then there is Intel. Granted they produce great chips. But the whole cpu market is under their control. They are releasing new chipsets and sockets for breakfast.
We need stronger competitors4 -
My new motherboard (MSI Z170 GAMING AC) reports temps for the memory, CPU, GPU and hard drive in °C. And the motherboard temp in °F.
Almost had a fucking heart attack the first time I opened Speccy for benchmarking. -
So recently I did a lot of research into the internals of Computers and CPUs.
And i'd like to share a result of mine.
First of all, take some time to look at the code down below. You see two assembler codes and two command lines.
The Assembler code is designed to test how the instructions "enter" and "leave" compare to manually doing what they are shortened to.
Enter and leave create a new Stackframe: this means, that they create a new temporary stack. The stack is where local variables are put to by the compiler. On the right side, you can see how I create my own stack by using
push rbp
mov rbp, rsp
sub rsp, 0
(I won't get into details behind why that works).
Okay. Why is this even relevant?
Well: there is the assumption that enter and leave are very slow. This is due to raw numbers:
In some paper I saw ( I couldn't find the link, i'm sorry), enter was said to use up 12 CPU cycles, while the manual stacking would require 3 (push + mov + sub => 1 + 1 + 1).
When I compile an empty function, I get pretty much what you'd expect just from the raw numbers of CPU cycles.
HOWEVER, then I add the dummy code in the middle:
mov eax, 123
add eax, 123543
mov ebx, 234
div ebx
and magically - both sides have the same result.
Why????
For one thing, there is CPU prefetching. This is the CPU loading in ram before its done executing the current instruction (this is how anti-debugger code works, btw. Might make another rant on that). Then there is the fact that the CPU usually starts work on the next instruction while the current instruction is processing IFF the register currently involved isnt involved in the next instruction (that would cause a lot of synchronisation problems). Now notice, that the CPU can't do any of that when manually entering and leaving. It can only start doing the mov eax, 1234 while performing the sub rsp, 0.
----------------
NOW: notice that the code on the right didn't take any precautions like making sure that the stack is big enough. If you sub too much stack at once, the stack will be exhausted, thats what we call a stack overflow. enter implements checks for that, and emits an interrupt if there is a SO (take this with a grain of salt, I couldn't find a resource backing this up). There are another type of checks I don't fully get (stack level checks) so I'd rather not make a fool of myself by writing about them.
Because of all those reasons I think that compilers should start using enter and leave again.
========
This post showed very well that bare numbers can often mislead.21 -
I do not like the direction laptop vendors are taking.
New laptops tend to feature fewer ports, making the user more dependent on adapters. Similarly to smartphones, this is a detrimental trend initiated by Apple and replicated by the rest of the pack.
As of 2022, many mid-range laptops feature just one USB-A port and one USB-C port, resembling Apple's toxic minimalism. In 2010, mid-class laptops commonly had three or four USB ports. I have even seen an MSi gaming laptop with six USB ports. Now, much of the edges is wasted "clean" space.
Sure, there are USB hubs, but those only work well with low-power devices. When attaching two external hard drives to transfer data between them, they might not be able to spin up due to insufficient power from the USB port or undervoltage caused by the impedance (resistance) of the USB cable between the laptop's USB port and hub. There are USB hubs which can be externally powered, but that means yet another wall adapter one has to carry.
Non-replaceable [shortest-lived component] mean difficult repairs and no more reserve batteries, as well as no extra-sized battery packs. When the battery expires, one might have to waste four hours on a repair shop for a replacement that would have taken a minute on a 2010 laptop.
The SD card slot is being replaced with inferior MicroSD or removed entirely. This is especially bad for photographers and videographers who would frequently plug memory cards into their laptop. SD cards are far more comfortable than MicroSD cards, and no, bulky external adapters that reserve the device's only USB port and protrude can not replace an integrated SD card slot.
Most mid-range laptops in the early 2010s also had a LAN port for immediate interference-free connection. That is now reserved for gaming-class / desknote laptops.
Obviously, components like RAM and storage are far more difficult to upgrade in more modern laptops, or not possible at all if soldered in.
Touch pads increasingly have the buttons underneath the touch surface rather than separate, meaning one has to be careful not to move the mouse while clicking. Otherwise, it could cause an unwanted drag-and-drop gesture. Some touch pads are smart enough to detect when a user intends to click, and lock the movement, but not all. A right-click drag-and-drop gesture might not be possible due to the finger on the button being registered as touch. Clicking with short tapping could be unreliable and sluggish. While one should have external peripherals anyway, one might not always have brought them with. The fallback input device is now even less comfortable.
Some laptop vendors include a sponge sheet that they want users to put between the keyboard and the screen before folding it, "to avoid damaging the screen", even though making it two millimetres thicker could do the same without relying on a sponge sheet. So they want me to carry that bulky thing everywhere around? How about no?
That's the irony. They wanted to make laptops lighter and slimmer, but that made them adapter- and sponge sheet-dependent, defeating the portability purpose.
Sure, the CPU performance has improved. Vendors proudly show off in their advertisements which generation of Intel Core they have this time. As if that is something users especially care about. Hoo-ray, generation 14 is now yet another 5% faster than the previous generation! But what is the benefit of that if I have to rely on annoying adapters to get the same work done that I could formerly do without those adapters?
Microsoft has also copied Apple in demanding internet connection before Windows 11 will set up. The setup screen says "You will need an Internet connection…" - no, technically I would not. What does technically stand in the way of Windows 11 setting up offline? After all, previous Windows versions like Windows 95 could do so 25 years earlier. But also far more recent versions. Thankfully, Linux distributions do not do that.
If "new" and "modern" mean more locked-in and less practical and difficult to repair, I would rather have "old" than "new".12 -
Got new hardware for my minecraft server!
Intel I7-8068K 6 Cores 12 Threads 5.1GHz
64GB DDR4 RAM 2666MHz
2x 512GB NVMe (RAID 1)
First time I have ever used a 5GHz CPU I am so excited!12 -
After long time logged into windows desktop and dared to open internet explorer then I seen option
"What's New in Internet Explorer 8"
So I thought let's see what's new in IE8
When I clicked on that option, suddenly my CPU fan stopped working.
There was no connection between them,
But what a coincidence!
F***U IE:D2 -
I waited 8 years for cyberpunk 2077. I also bought the most expensive pc of my life including 34inch ultra wide monitor.
But at the last moment I got fucked by AMD. The 5900x that I ordered was dead on arrival and went back for replacement. But they don't have any stock to replacement. So I am fucking waiting for cpu. They are not providing any information on the stock.
I am getting depressed day by day.21 -
I'm the only one working on this anymore and every toolchain supporting the system (remember, we're using an ARM9 [initial strap CPU] AND an ARM11 (give or take an ARM7 slaved to the ARM9 that we don't have support for yet), all in tandem, and the only toolchain that remotely works is for ARM6 for some reason) hates the Linux kernel. Current goals: SD R/W support (currently RO), X, GNUTools, maybe a better fucking softkey driver (i'll have to find whoever made this one and fucking beat him), and a working joy2mouse/touch2mouse driver. Oh, and figure out if Swap would work either with the New 2DS/3DS' Bonus Drive (unused 64MB partition on NAND) without killing the NAND as the SD access is max. 1.2 MB/s read/write speed or so, which isn't fast enough for swap AND other things.
Currently working:
Busybox
Read-only SD support
Weston (term only, can't click)
Standard 3DS/Standard 2DS/New 3DS (Models before 2017, the non-foldables, rebranded standard 2DSes) features only, not yet New 3DS/New 2DS-enhanced
Currently failing final compile because toolchain:
Preliminary custom R/W SD support4 -
After a few weeks of being insanely busy, I decided to log onto Steam and maybe relax with a few people and play some games. I enjoy playing a few sandbox games and do freelance development for those games (Anywhere from a simple script to a full on server setup) on the side. It just so happened that I had an 'urgent' request from one of my old staff member from an old community I use to own. This staff member decided to run his own community after I sold mine off since I didn't have the passion anymore to deal with the community on a daily basis.
O: Owner (Former staff member/friend)
D: Other Dev
O: Hey, I need urgent help man! Got a few things developed for my server, and now the server won't stay stable and crashes randomly. I really need help, my developer can't figure it out.
Me: Uhm, sure. Just remember, if it's small I'll do it for free since you're an old friend, but if it's a bigger issue or needs a full recode or whatever, you're gonna have to pay. Another option is, I tell you what's wrong and you can have your developer fix it.
O: Sounds good, I'll give you owner access to everything so you can check it out.
Me: Sounds good
*An hour passes by*
O: Sorry it took so long, had to deal with some crap. *Insert credentials, etc*
Me: Ok, give me a few minutes to do some basic tests. What was that new feature or whatever you added?
O: *Explains long feature, and where it's located*
Me: *Begins to review the files* *Internal rage wondering what fucking developer could code such trash* *Tests a few methods, and watches CPU/RAM and an internal graph for usage*
Me: Who coded this module?
O: My developer.
Me: *Calm tone, with a mix of some anger* So, you know what, I'm just gonna do some simple math for ya. You're running 33 ticks a second for the server, with an average of about 40ish players. 33x60 = 1980 cycles a minute, now lets times that by the 40 players on average, you have 79,200 cycles per minute or nearly 4.8 fucking cycles an hour (If you maxed the server at 64 players, it's going to run an amazing fucking 7.6 million cycles an hour, like holy fuck). You're also running a MySQLite query every cycle while transferring useless data to the server, you're clusterfucking the server and overloading it for no fucking reason and that's why you're crashing it. Another question, who the fuck wrote the security of this? I can literally send commands to the server with this insecure method and delete all of your files... If you actually want your fucking server stable and secure, I'm gonna have to recode this entire module to reduce your developer's clusterfuck of 4.8 million cycles to about 400 every hour... it's gonna be $50.
D: *Angered* You're wrong, this is the best way to do it, I did stress testing! *Insert other defensive comments* You're just a shitty developer (This one got me)
Me: *Calm* You're calling me a shitty developer? You're the person that doesn't understand a timer, I get that you're new to this world, but reading the wiki or even using the game's forums would've ripped this code to shreds and you to shreds. You're not even a developer, cause most of this is so disorganized it looks like you copy and pasted it. *Get's angered here and starts some light screaming* You're wasting CPU usage, the game can't use more than 1 physical core, and after a quick test, you're stupid 'amazing' module is using about 40% of the CPU. You need to fucking realize the 40ish average players, use less than this... THEY SHOULD BE MORE INTENSIVE THAN YOUR CODE, NOT THE OPPOSITE.
O: Hey don't be rude to Venom, he's an amazing coder. You're still new, you don't know as much as him. Ok, I'll pay you the money to get it recoded.
Me: Sounds good. *Angered tone* Also you developer boy, learn to listen to feedback and maybe learn to improve your shitty code. Cause you'll never go anywhere if you don't even understand who bad this garbage is, and that you can't even use the fucking wiki for this game. The only fucking way you're gonna improve is to use some of my suggestions.
D: *Leaves call without saying anything*
TL;DR: Shitty developer ran some shitty XP system code for a game nearly 4.8 million times an hour (average) or just above 7.6 million times an hour (if maxed), plus running MySQLite when it could've been done within about like 400 an hour at max. Tried calling me a shitty developer, and got sorta yelled at while I was trying to keep calm.
Still pissed he tried calling me a shitty developer... -
>Gets a new CPU for desktop (yay, went from R5 1600 to R5 3600X)
>Spends half a day flashing new MB BIOS (Needed to flash individual major versions in order, couldn't just go 1.10 to 6.40)
>Finally finishes preparations and goes to replace the CPU
>Cleans the old one and packages it to give it to a friend
>Has issues inserting the new one as the orientation arrow on the motherboard was very hard to make out
>Spends 30 minutes applying thermal paste, worrying about optimal spread
>Forgets which side the CPU fan goes on
>Finally boots back up... CPU fan is suddenly loud AF under load, but eh, temps under stress are sub-60, so, good
~~Next day~~
>Loud CPU fan is too annoying, opens the case again
>CPU fan is on backwards
Ugh
>Takes the fan off, turns it around and fastens again, puts PC back together and boots
>Is quiet again, nice
>Goes to work on the PC
>2 hours later randomly checks temps because no fan noise is weird
>CPU at 75dC, crap
>Opens the (live) PC, CPU fan is not spinning
>Has put the header on one pin to a side
>Unplugs and replugs it correctly
>Fan suddenly starts spinning very fast and cuts my finger
>Finally closes the case once more. All issues resolved
...Its situations like these that make me wonder... What would happen it I had to work with servers in person, physically lol8 -
Follow-up to my previous story: https://devrant.com/rants/1969484/...
If this seems to long to read, skip to the parts that interest you.
~ Background ~
Maybe you know TeamSpeak, it's basically a program to talk with other people on servers. In TeamSpeak you can generate identities, every identity has a security level. On your server you can set a minimum security level you need to connect. Upgrading the security level takes longer as the level goes up.
~ Technical background ~
The security level is computed by doing this:
SHA1(public_key + offset)
Where public_key is your public key in Base64 and offset is an 8 Byte unsigned long. Offset is incremented and the whole thing is hashed again. The security level comes from the amount of Zero-Bits at the beginning of the resulting hash.
My plan was to use my GPU to do this, because I heared GPUs are good at hashing. And now, I got it to work.
~ How I did it ~
I am using a start offset of 0, create 255 Threads on my GPU (apparently more are not possible) and let them compute those hashes. Then I increment the offset in every thread by 255. The GPU also does the job of counting the Zero-Bits, when there are more than 30 Zero-Bits I print the amount plus the offset to the console.
~ The speed ~
Well, speed was the reason I started this. It's faster than my CPU for sure. It takes about 2 minutes and 40 seconds to compute 2.55 Billion hashes which comes down to ~16 Million hashes per second.
Is this speed an expected result, is it slow or fast? I don't know, but for my needs, it is fucking fast!
~ What I learned from this ~
I come from a Java background and just recently started C/C++/C#. Which means this was a pretty hard challenge, since OpenCL uses C99 (I think?). CUDA sadly didn't work on my machine because I have an unsupported GPU (NVIDIA GeForce GTX 1050 Ti). I learned not to execute an endless loop on my GPU, and so much more about C in general. Though it was small, it was an amazing project.1 -
Finally bought a macbook for my android app development. Paid 1600 EUR for 2015 MID mbpro with 512GB SSD, 16gb ram, 2.5ghz quad core i7 cpu and dedicated gpu. I know I overpaid but I couldnt fkin stand working on windows anymore. I spent one year with fcking windows building android apps. Making apps on windows is fking impossible. And dont even get me started on new crap macbooks starting from 2016 :/19
-
Incredibly drunk last night.
Decided to take the CD/DVD drive out of my old PC (now a server) and put it into my new gaming PC.
Required a partial rebuild because:
1) GPU was too close to PSU to plug in a new SATA power line
2) CPU cooling lines were directly in the way of the space for the DVD/CD bay
3) Had to route the new power cables of course
Incredibly, even though I couldn't stand, I was able to perfectly execute the maintenance on both machines and I'm even using the new PC to write this rant now.
I must be a God.5 -
Got my new PC up and running!
CPU: AMD Ryzen 5 3400G (APU)
Cooler: Scythe Mugen 5
Mobo: Asus B450-F Gaming ROG Strix
RAM: Crucial Ballistix 2x16GB
SSD: Samsung 2TB M2 Evo 860
DVD: Plextor PX-891SAF
PSU: Bequiet Straight Power 11 550W
Case: Lancool PC-K58 (10 years old)
Case fans: Bequiet Silentwings 3 140mm (front), Silentwings 2 120mm (back)
The cooler is massively oversized for the CPU, but perfect for a silent PC.
OS: Linux Mint 20 Cinnamon
As much as I loved Win 7, but it's over, and Win 10 just isn't acceptable.18 -
I haven't built a new computer in 9 years, but I finally, finally got together enough money to go big.
POWER: Seasonic Focus Plus 850W 80+ Gold
MOTHERBOARD: NZXT N7 Z370 ATX LGA1151 (Black)
GPU: EVGA GTX 1080 FTW GAMING
CPU: INTEL Core i7-8700K 3.7GHz 6-Core
RAM: G.SKILL TridentZ 3000 Mhz 32GB (2 x 16 GB)
SSD: SAMSUNG 860 EVO 1TB
CPU COOLING: NZXT Kraken X62
I can't fucking wait for it to arrive.5 -
So not too long ago I made a rant about this time I told my superiors of a problem and it got shot down because "We don't want to save developer time".
I warned them that the CPU was going to reach it's limit because the issue involves exponential growth on CPU usage the more we develop.
As it turned out, despite my thoughts that we wouldn't reach that limit for over a year, a new development led to it blowing up in their face today.
I am now treated as the expert on the topic and they're rushing to plan my suggested implementation in the next release.
I'm mildly amused.5 -
New cooler for CPU installed.
Damn it. Was long overdue, old cooler was way undersized after an CPU upgrade.
I knew it would be tricky. But damn... My motherfugging stubbornness.
*sigh*
1mm over RAM heatspreader, took a long time getting blindly the cooler fixed... Screwing blindly isn't fun.13 -
Been so long, sup guys. bringing you good news (at least for me).
just got accepted on my new job. the funny thing is. i legit put "Made 16 bit CPU in Minecraft" on my resume. LMAO6 -
Black Friday is the time where you consider to buy a new laptop and finds that laptop companies don't know how to make one for programmers. Either they are shit or they are gamer laptop that have 2~4 GB of video memory you will never use and costs thousands of dollars.
Is it hard to make one with good CPU, RAM and a SSD without concluding that you are a hardcore gamer and you need an expensive video card along with it?8 -
Can someone help me understand?
I subscribed to a nifty IT-releated magazine, and on its back, there's an ad for "Dedicated root server hosting", nothing unusual at a first glance, but after I read the issue, I decided to humor them and see what it is that they offered, and... It just... Doesn't make sense to me!
An ad for "Dedicated Root Server" - What is a dedicated root server first of all? Root servers of any infrastructure sound pretty important.
But, the ad also boasts "High speed performance with the new Intel Core i9-9900K octa-core processor", that's the first weird thing.
Why would anyone responsible enough want to put an i9 into a highly-reliable root server, when the thing doesn't even support ECC? Also, come on, octa-core isn't much, I deal with servers that have anywhere between 2 and 24 cores. 8 isn't exactly a win, even if it has a higher per-core clock.
Oh, also, further down the ad has a list of, seeming, advantages/specs of the servers, they proclaim that the CPU "incl. Hyper-Threading-Technology"... Isn't that... Standard when it comes to servers? I have never seen a server without hyperthreading so far at my job.
"64 GBs of DDR4 RAM" - Fair enough, 64 gigs is a good amount, but... Again, its not ECC, something I would never put into a server.
"2 x 8 TB SATA Enterprise Hard Drive 7200 rpm" - Heh, "enterprise hard drive", another cheap marketing word, would impress me more if they mentioned an actual brand/model, but I'll bite, and say that at least the 7200 rpm is better than I expected.
"100 GBs of Backup Space" - That's... Really, really little. I've dealt with clients who's single database backup is larger than that. Especially with 2x8 TB HDD (Even accounting for software raids on top)
This one cracks me up - "Traffic unlimited"
Whaaaat?! You are not gonna give me a limit to the total transferred traffic to the internet for my server in your data center? Oh, how generous of you, only, the other case would make the server just an expensive paperweight! I thought this ad was for semi-professionals at least, so why mention traffic, and not bandwidth, the thing that matters much more when it comes to servers? How big of a bandwidth do I get? Don't tell me you use dialup for your "Dedicated Root Server"s!
"Location Germany or Finland" - Fair enough, geolocation can matter when it comes to latency.
"No minimum contract" - Oooh, how kiiiind of you, again, you are not gonna charge me extra for using the server only as long as I pay? How nice!
"Setup Fee £60" - I guess, fair enough, the server is not gonna set itself up, only...
The whole ad is for "monthly from £55.50", that's quite the large fee for setup.
Oh, and a cherry on top, the tiny print on the bottom mentions: "All prices exclude VAT and are a subject to..." blah blah blah.
Really? I thought that this sort of almost customer deceipt is present only in the common people's sphere!
I must say, there's being unimpressed, and then... There's this. Why, just... Why? Anyone understands this? Because I don't...12 -
I tried LSD yesterday!
Backstory: I have a weird combo of bipolar type 1 and autism. During the day, my brain works inconsistently. Here how my day usually goes:
09:00. I wake up. Uninterested, cold, masculine. No thoughts in the background. No OCD.
12:00. Brain warms up. Thought process begins. Thoughts are short in their length
14:00. Thoughts start to get longer. Stress starts to accumulate. Background thoughts start, now typically 2–3 at a time.
16:00. Twitching begins. Thought chains are now 5–6 concepts long, one following the other. Perception level rises quickly. I start to feel more feminine. It is in this state that I start to spot imperfections and mistakes looking at code or text without reading it. I see it like a painting, and mistakes appear as “visually wrong” parts. This does not depend on formatting.
17:00. OCD becomes more severe. I HAVE to touch all the surfaces around me, evenly, as if my hands were text highlighters, and I had to paint everything evenly, without overlaps or spots that are brighter or darker than the others. Some surface textures become irritating, and feel quite unpleasant to the touch. If I go for a run now, like 3 km or so, I feel somewhat relieved.
18:00. Things are getting serious. Creativity levels through the roof. I speak in long, never-ending, profound sentences. Background and foreground thoughts almost become one. I appear visually drunk and happy, despite never drinking alcohol. Femininity rises even further. Sometimes, when I speak to a small group of people, especially if I go with friends to meet new people, and we go to some bar, new people ask to record my voice or to write down whatever I’m saying. To be honest, this reason alone is a huge boost to how I see yourself.
19:00. OCD is crazy now. Surfaces have soul.
21:00 <— Gotta take my meds and go to sleep here to prevent what comes at 22:00
22:00. All thoughts, both foreground and background, fully became one. Now my brain officially disobeys me and thinks on its own, and I can ride it like a surfer at best. Twitching becomes concerning. I develop a 1000-yard stare. I am officially a female. Physical strength is somewhat enhanced. Pain tolerance lowered significantly.
23:00. Derealization begins. The world around me appears two-dimensional and flat, like a picture. It is hard to get home on foot, even in close (less than one km) proximity. Brain is fully numb. All that thought monstrosity that was building up is just noise now. Zero “flops” available to think about something I want to think about, like how much money I have on me or what time it is.
I go to sleep. I see nightmares. I wake up, and the cycle repeats.
Contrary to a popular opinion, I never take any “brain-boosting” meds like antidepressants, and I think now you can see why. I consume neither alcohol nor caffeine. Neither me, nor my doctors want my brain to explode. I only take lamotrigine that helps to “lower down” mania, and quetiapine, a neuroleptic, that slows down my brain, like a neuroleptic. Both are there to slow down my brain, to kinda “throttle” my brain like a CPU to cool it down.
That said, 100ug of LSD just… brought me my usual 18:00 state, but in the morning?
All that small-dose recreational ordeal? The thing that helps people feel more energetic and creative?
People pay money… for that? To feel the way I feel every evening?10 -
EDIT: !rant, forgot tags, sorry :(
Tfw your new laptop CPU gets the same performance as your 10y.o. desktop and is 10x more power efficient.
Ryzen 5 2500U (2018):
> 4c/8t
> 2GHz base / 3.6 turbo
> 15W TDP
Xeon W3520 (2009):
> 4c/8t
> 2.7GHz base / 2.9 turbo
>130W TDP
Running new hardware feels great.3 -
(long post is long)
This one is for the .net folks. After evaluating the technology top to bottom and even reimplementing several examples I commonly use for smoke testing new technology, I'm just going to call it:
Blazor is the next Silverlight.
It's just beyond the pale in terms of being architecturally flawed, and yet they're rushing it out as hard as possible to coincide with the .Net 5 rebranding silo extravaganza. We are officially entering round 3 of "sacrifice .Net on the altar of enterprise comfort." Get excited.
Since we've arrived here, I can only assume the Asp.net Ajax fiasco is far enough in the past that a new generation of devs doesn't recall its inherent catastrophic weaknesses. The architecture was this:
1. Create a component as a "WebUserControl"
2. Any time a bound DOM operation occurs from user interaction, send a payload back to the server
3. The server runs the code to process the event; it spits back more HTML
Some client-side js then dutifully updates the UI by unceremoniously stuffing the markup into an element's innerHTML property like so much sausage.
If you understand that, you've adequately understood how Blazor works. There's some optimization like signalR WebSockets for update streaming (the first and only time most blazor devs will ever use WebSockets, I even see developers claiming that they're "using SignalR, Idserver4, gRPC, etc." because the template seeds it for them. The hubris.), but that's the gist. The astute viewer will have noticed a few things here, including the disconnect between repaints, inability to blend update operations and transitions, and the potential for absolutely obliterative, connection-volatile, abusive transactional logic flying back and forth to the server. It's the bring out your dead approach to seeing how much of your IT budget is dedicated to paying for bandwidth and CPU time.
Blazor goes a step further in the server-side render scenario and sends every DOM event it binds to the server for processing. These include millisecond-scale events like scroll, which, at least according to GitHub issues, devs are quickly realizing requires debouncing, though they aren't quite sure how to accomplish that. Since this immediately becomes an issue with tickets saying things like, "scroll event crater server, Ugg need help! You said Blazorclub good. Ugg believe, Ugg wants reparations!" the team chooses a great answer to many problems for the wrong reasons:
gRPC
For those who aren't familiar, gRPC has a substantial amount of compression primarily courtesy of a rather excellent binary format developed by Google. Who needs the Quickie Mart, or indeed a sound markup delivery and view strategy when you can compress the shit out of the payload and ignore the problem. (Shhh, I hear you back there, no spoilers. What will happen when even that compression ceases to cut it, indeed). One might look at all this inductive-reasoning-as-development and ask themselves, "butwai?!" The reason is that the server-side story is just a way to buy time to flesh out the even more fundamentally broken browser-side story. To explain that, we need a little perspective.
The relationship between Microsoft and it's enterprise customers is your typical mutually abusive co-dependent relationship. Microsoft goes through phases of tacit disinterest, where it virtually ignores them. And rightly so, the enterprise customers tend to be weaksauce, mono-platform, mono-language types who come to work, collect a paycheck, and go home. They want to suckle on the teat of the vendor that enables them to get a plug and play experience for delivering their internal systems.
And that's fine. But it's also dull; it's the spouse that lets themselves go, it's the girlfriend in the distracted boyfriend meme. Those aren't the people who keep your platform relevant and competitive. For Microsoft, that crowd has always been the exploratory end of the developer community: alt.net, and more recently, the dotnet core community (StackOverflow 2020's most loved platform, for the haters). Alt.net seeded every competitive advantage the dotnet ecosystem has, and dotnet core capitalized on. Like DI? You're welcome. Are you enjoying MVC? Your gratitude is understood. Cool serializers, gRPC/protobuff, 1st class APIs, metadata-driven clients, code generation, micro ORMs, etc., etc., et al. Dear enterpriseur, you are fucking welcome.
Anyways, b2blazor. So, the front end (Blazor WebAssembly) story begins with the average enterprise FOMO. When enterprises get FOMO, they start to Karen/Kevin super hard, slinging around money, privilege, premiere support tickets, etc. until Microsoft, the distracted boyfriend, eventually turns back and says, "sorry babe, wut was that?" You know, shit like managers unironically looking at cloud reps and demanding to know if "you can handle our load!" Meanwhile, any actual engineer hides under the table facepalming and trying not to die from embarrassment.36 -
!dev
Guys, we need talk raw performance for a second.
Fair disclaimer - if you are for some reason intel worker, you may feel offended.
I have one fucking question.
What's the point of fucking ultra-low-power-extreme-potato CPUs like intel atoms?
Okay, okay. Power usage. Sure. So that's one.
Now tell me, why in the fucking world anyone would prefer to wait 5-10 times more for same action to happen while indeed consuming also 5-10 times less power?
Can't you just tune down "big" core and call it a day? It would be around.. a fuckton faster. I have my i7-7820HK cpu and if I dial it back to 1.2Ghz my WINDOWS with around lot of background tasks machine works fucking faster than atom-powered freaking LUBUNTU that has only firefox open.
tested i7-7820hk vs atom-x5-z8350.
opening new tab and navigating to google took on my i7 machine a under 1 second, and atom took almost 1.5 second. While having higher clock (turbo boost)
Guys, 7820hk dialled down to 1.2 ghz; 0.81v
Seriously.
I felt everything was lagging. but OS was much more responsive than atom machine...
What the fuck, Intel. It's pointless. I think I'm not only one who would gladly pay a little bit more for such difference.
i7 had clear disadvantages here, linux vs windows, clear background vs quite a few processes in background, and it had higher f***ng clock speed.
TL;DR
Intel atom processors use less power but waste a lot of time, while a little bit more power used on bigger cpu would complete task faster, thus atoms are just plain pointless garbage.
PS.
Tested in frustration at work, apparently they bought 3 craptops for presentations or some shit like that and they have mental problems becouse cheapest shit on market is more shitty than they anticipated ;-;
fucking seriously ;-;16 -
FUUCCKKKK!! I need to hit smth. Or rant..
So that flaky ec2 issue.. These ec2s act as a shared environment for multiple apps. Our app is one of them. I have no access to those ec2s at all.
What I have access to is my app and some monitoring. Now the app randomly starts lagging while nearly idling. At the same random times monitoring stops completely and doesn't come back up. This happens to random app instances at random times.
Reached out to infra support, managed to get attention from the big boys [mgmt]. Today we got the fix deployed. I test it out -- problem persists.
I find this behaviour somewhat familiar. Managed to get some server stats from infra folks. Apparently cpu% is high as well as load avg [cpu queue]. Bingo! I know how to fix it!
So I write a long comment w/ all the commands and all the 'if that, do this'. Send it to one of the infra technitians
and I get a reply: 'we will apply cpu usage limitations to fix the issue'
wait... Cpu% limitations will do nothing but highlight the underlying problem...
'no, instances have high cpu utilisation which is causing those lags. We will limit cpu resources and it will be fixed'
oh ffs... Cpu utilization and cpu queue are VERY different things.. I tried explaining that to them like 7-9 times. And all I get is:
'yes, cpu utilization is the problem. We will limit it and solve the problem'
I would surely escalate all of this through higher channels if only I could get my hands on those ec2s and have a proof. But that is not happening and I'm forced to sit back and watch them break things even worse until they are out of options and mark my query as 'wont fix'....
Fuck that's frustrating....
*thinking to myself* so I've read about that new vulnerability 2 days ago that allows one to escape from docker container to the host... What if <...>4 -
At the institute I did my PhD everyone had to take some role apart from research to keep the infrastructure running. My part was admin for the Linux workstations and supporting the admin of the calculation cluster we had (about 11 machines with 8 cores each... hot shit at the time).
At some point the university had some euros of budget left that had to be spent so the institute decided to buy a shiny new NAS system for the cluster.
I wasn't really involved with the stuff, I was just the replacement admin so everything was handled by the main admin.
A few months on and the cluster starts behaving ... weird. Huge CPU loads, lots of network traffic. No one really knows what's going on. At some point I discover a process on one of the compute nodes that apparently receives commands from an IRC server in the UK... OK code red, we've been hacked.
First thing we needed to find out was how they had broken in, so we looked at the logs of the compute nodes. There was nothing obvious, but the fact that each compute node had its own public IP address and was reachable from all over the world certainly didn't help.
A few hours of poking around not really knowing what I'm looking for, I resort to a TCPDUMP to find whether there is any actor on the network that I might have overlooked. And indeed I found an IP adress that I couldn't match with any of the machines.
Long story short: It was the new NAS box. Our main admin didn't care about the new box, because it was set up by an external company. The guy from the external company didn't care, because he thought he was working on a compute cluster that is sealed off behind some uber-restrictive firewall.
So our shiny new NAS system, filled to the brink with confidential research data, (and also as it turns out a lot of login credentials) was sitting there with its quaint little default config and a DHCP-assigned public IP adress, waiting for the next best rookie hacker to try U:admin/P:admin to take it over.
Looking back this could have gotten a lot worse and we were extremely lucky that these guys either didn't know what they had there or didn't care. -
Hmmmm so the MacBook I'm looking at just got upgraded with a new CPU and better screen... Hmmmm if only the 16gb of RAM didn't cost so much!
Still think I might go for it, general consensus seems that it's the better option over a pixelbook10 -
Upgraded from a gen 6 motherboard and cpu to a gen 11. And the difference is asounding.
My old setup did everything I needed it to pretty well. Hell it got me through college as well as multiple part time and full time jobs. As well as the system that I learned about 80% of what I know specifically for my career
But alas the old cpu's age started showing and I'd consistently throttle multiple threads when compiling or even just using 7z on large folders. Hell deleting anything above a couple hundred mb would pause for a few seconds
The new cpu is immensely faster on such tasks (probably cause of the extra 2 cores and cpu improvements since gen 6)
I haven't been this excited since I got my first SSD a few years ago and saw the immense boot time improvements. Maybe soon I'll get an nvme drive and really be zooming with this setup 🤣🤣🤣6 -
In 11th grade I had a shitty teacher who would make students go away from this one computer he always used in the computer lab. It was pretty annoying because he only ever used it to watch YouTube videos of fuckall Bollywood news and other garbage like it.
It was extremely annoying and inconvenient to everyone.
One day, I went to his PC, when he was not there, and reallocated the default way to open all .exe files to nothing, using the Windows Registry.
He came back and saw that none of his applications work and his PC is basically unusable now.
It was extremely funny for a few days to watch the IT guys try to fix but they weren't able to.
Now I don't know what they told the school but a month later they bought a completely new CPU and placed it instead of that one.
So fucking stupid.2 -
Tl; dr: Linux on Ryzen is a pain at the moment.
Now for the long part: Our student council got new computers because the old ones where slow as hell. As one of the admins, the others and I together decided that ryzen would be a good option, because they are not that expensive and we wouldn't have to buy gpus. (Wrong decision it turns out.) We settled on the ryzen 3 2200G and bought three systems to replace the old ones.
We meet Saturday morning and build the systems. All was fine and we were happy. The we tried to install ubuntu via preseeded netboot, which seemed to work fine at first. Then we started having weird screen issues and couldn't proceed with the installation. (See image) we then grumpily decided to just install them all one by one, flashed two usbs and started installing. On two systems the installation worked and we installed our packages, we weren't so lucky with the third one. It would crash on us all the time, even in bios. While that was going on we tried to set the other two up, turns out those two were also crashing but not as frequent as the other one. So we start to google and find people saying that kernel 4.19 kinda fixes it. We install it on the two working machines and the crashes get less frequent but are still there. At that point it was midnight and we went home.
Sunday morning: we reseated the cpu on the third system and it seems to be better now (it installed on the second try) and we were able to change the kernel. Yay. Now all three are in a state where they will sometimes randomly reset. :/ and we don't know what to try anymore.... Any suggestions?1 -
Im excited because I'm getting new pc parts for free.
The parts are:
* AMD 8 core CPU ( not sure which model )
* 16 GB ram
* some good graphics card ( not sure what model )
* motherboard
* cooling ( still not Sure what it will be, the guy was very vague )
It's also coming just before my birthday7 -
Got a new laptop, but it was not properly working, so while waiting for it to get replaced, I have to use the old one again.
Feels like the CPU is a freaking potato!2 -
Working with acutal BigData. Will be ''promoted'' to a new team where I will work on a system wrote in php+mysql with literal millions of requests and database rows. We are currently seeing server crashes around once a week on peak usage. Stack is a vps 64gb ram server + i dont know how many cpu/ cores. Apache, php, mysql.Best ways to optimise and adapt in this case? Kafka? Rabbitmq?ngnix? More hardware?21
-
I was flash developer once, it was great when macromedia was around, then adobe acquired them, now flash is gone.
Years are passing and most of industry is the same as always. Trying to drag you into this rat race of learning new amazing technologies, amazing projects that are actually doing same job as 50 years ago but using more memory and cpu cycles. Because all has it’s roots in algorithms from previous centuries.
So youngsters loose your best life time, be innovative by doing nothing more then copy paste from stackoverflow and duck typing shitty code.
Be a slave and sit in the amazing office, that has everything but not your real life that meanwhile is sucked by corporate squeezer till your last breath.
Be piece of shit that can be kicked around.
Watch youtube, facebook, instagram or whatever social network that shows you pictures that are fooling your mind that you’re someone special and you need this stuff.
Then be ready to suck some dicks to earn money and buy stuff you don’t need, live where you don’t want and do what you don’t like. You piece of shit.
Well that’s what disappoints me from my tech stack.
Now chill out, turn off your electronic gadgets, go out and enjoy real world.1 -
This always gets me:
Developers complaining that their 4 year old / cheap ass computer is slow.
Get. A. New. One.
It's not that hard.
Here, let me do one for you:
https://computeruniverse.net/en/...
I just went to a site that delivers across Europe, and selected a cheap laptop with a decent CPU and SSD. Short on RAM, sure, and without a Windows License. But you can buy RAM for an additional 50$, and that brings you to a total of 550€, delivery included. And it will WORK. And it will be fast.
It's too expensive?
No, not exactly. Wherever you are in the world, if you can code decently, good enough to have the right to complain about development tools, you are eligible to at least 10$ per hour income as a freelancer across the globe. I've had such opportunities offered to me by many organizations, especially non-profit ones that need cheap employees. I actually was offered more but let's stick to 10$ per hour.
So that's 1600$ per month. Enough to buy 3 such laptops. Oh, taxes, I forgot. So you get 2 laptops. Wait! You need food and everything else. Well if you're in a country where that offer actually makes sense, then it's likely that you can live off of 400$ per month quite well. Maybe 800$ if you need to pay rent.
So that's roughly 1 month of work for a laptop that will make you not waste time on waiting for stuff.
Sweet! 1 Month! What does it get me?
Well assuming that you have no laptop, it gets you A JOB that pays you 1600$ per month.
But if you DO have a laptop, you can sell it for cheap, and benefit from the following:
1. Boot-up time from 30-60 seconds to 10 seconds.
2. Installing software - from 1 minute to 10 seconds.
3. Opening a browser - from 10 seconds to 1 second.
4. Opening an advanced text editor (Atom, VS.Code) - from 10 seconds to 1 second.
5. Searching for a file on your entire hard drive - from 1 hour to 2 minutes.
....
You get the point. Waiting is reduced by several times.
So how much do you really wait when coding?
Well are you compiling? Are you opening a new project and the IDE needs to re-index the files? Are you opening programs like a terminal emulator, browser and such? Are you using virtual machines for dev environments?
Well all of these processes become several times faster. Depending on how often you do it, you'll be saving yourself from 1 hour per day to upto 4 hours per day (my case, where a HDD would be just out of the question).
How much is that time worth? At least 10$ per day. If you're working for 20 days per month, 240 days per year, that's a total of 2400$. And for the life time of that crappy laptop of 2 years, that's 4800$ saved. And that's with hugely conservative numbers. Nobody pays 10$ per hour any more, except if you've just started in the industry. I know because I've been there.
Please, for all that's sacred to you, justify right here, right now, HOW THE FUCK can you not afford to get that 8GB of RAM, that cheap ass SSD for 100$, or even a brand new laptop (hey! it's even portable and has FHD graphics on it!) for 550$.
That's why every time I hear someone who is a professional developer complain that they don't have money for a decent machine, I have to ask: why the fuck are you wasting yours and everyone else's time?!10 -
Why would anyone use CPU ticks for connection timeouts? Everytime that software is used in a new CPU it gets messed up. Why not just milliseconds?
-
At work we get points each month if we don't miss work for the entire month. I have enough points to get a Canon EOS eSLR, or a new mobo-cpu-mem combo. I hate having to make desicions and choices.6
-
What was supposed to be a simple HDD to SSD upgrade for my brother ended up taking 3 fucking days.
His computer's idle temp was fucking 90c because of fucking YEARS worth of dust preventing fans from breathing, one of them was even dead so I had to go to best buy to get a new one.
Had a power outage happen in the middle of cleaning the damn dust out before replacing the bad fan. It took close to 24 hours for it to come back, and I just didn't bother until the day after cause fuck that.
When I go back to it, I finish cleaning the dust, clone the old HDD to the new SSD, do the swap out, replace the thermal paste on CPU cause why the fuck not at this point. Then turn it on to test, yeah from 90c down to 20c : |
Yeah... so I charged him for the new fan, and the hour-ish it took me to clean the damn dust and replace the fan... (drive was agreed to be done for free as a favor) Yeah not really worth charging for but I was pissed and that was the 'nicest' way I could show how annoyed I was with him not properly caring for the PC. Like honestly 90 degrees celsius. I refuse to believe he didn't notice impacted performance in his games7 -
iOS is rotting my soul.
I've been a user of iPhone for 6 years now. For the first couple years, I wasnt really mindful of software I use, or I guess I didnt really care. As long as it did the bare minimum, I.e. bank app, call, text, browse, watch youtube vids, I didnt really care. However, in the last couple years, ive become very interested in tech and have worked on small developer projects, spent a lot of time coding in my free time, found really inspiring software and apps on my regular computer that just blow my mind on how advanced they are, and how I, some dumb guy with internet access, can just download it on my PC and use it.
This led me into a kind of software honeymoon phase, where I created a shiny new Github account and started exploring what other cool tools are just out there, available to me for free. My software honeymoon was spent on the beaches and resorts of the open-source software ecosystem. Exploring the gem-bearing caves and beautiful forests of anything from free open-source OCR programs(I needed it to convert my dads manuscript from scanned PDF .jpeg's to actual UTF8 text) to open-source RGB lighting/keymapping software to escape the memory-and-CPU-hungry(and most likely advertising-ID-interested) proprietary software that comes with the brand of mouse/keyboard/controller/etc.
It was like I was a kid exploring Disneyland for the first time or something. But then... then... I got off my computer. Picked up my phone to check notifications. Ew, tinder is blowing up notification center with marketing shit. I go to settings. Notification settings. Tinder's at the bottom so I just want to use a search bar instead of scrolling. There's no search bar. Minor inconvenience. Dark mode isnt dark enough for me. I guess thats just too damn bad, because for the next two hours, I'll have to figure it out by messing with accessibility settings. Time for bed, and I'm just getting plum tired of having to turn on my alarms every night for work the next morning. So I used the 'Automations' app to do it for me. For the next two weeks, at the time specified, 'There was an error running your automation' until I just delete the automation. Browsing through the FaceID settings, I see 'Attention Aware Features'. Cool, maybe now my phone won't automatically dim the screen when im in the middle of reading notifications on my lock screen. Haha, nope still does it. After turning on my alarms, I go to sleep. I wake up an hour late for work because those handy 'Attention Aware Features' silenced my alarm immediately because I fell asleep watching a youtube video.
I could go on and on. Its actually making me feel depressed typing this on my phone, fighting with Apple's primitive autocorrect and annoying implementation of Swype to type.4 -
when KhronosGroup anounced Vulkan back then, they also announced a whole set of software, that can handle all the new formats, that they introduced.
One format in particular peaked my interest recently, which is ktx2. It's an image format, that can be multilayered, and supercompressed, has inline mipmapping, and most importantly: streamed directly to the GPU, without involving the CPU basically at all.
Now here comes the kicker. If i want to use this format (mind you: Vulkan is around for a while now) for creating Skyboxes, there is only a single tool, that can properly convert hdr images to ktx2, and it only works on windows. Oh and there are no binaries, so in every case you have to compile it yourself.
Ah and then i thought, okay what if i then already render the cubemap faces and assemble them by hand into the cubemap, because _some_ ktx tools work on linux, then that should work right? wrong. When assembling it, it turns out, that now it's a 2D image instead of a 2DArray image with one element (which apparently is not the same for skyboxes)
Why is this shit such a pain in the ass?
Like.. I'm currently rendering equirectangular hdr images on my linux machine, then move these (usually 100MB) files over to some windows PC, convert it there into ktx2 cubemaps and then move it back. And everytime i need to do a change on the skybox, i have to repeat this whole nonsense. Ah.. and this tool doesn't even properly work on Windows, like you can't just disable mipmaps or change the filtering, because then the skybox is just black for some reason.
The funniest thing is, at the end of the day, these ktx2 files work on linux, as well as windows, mac and even mobile platform, so there's really no reason, that the conversion tool only works on one of them systems.
But hey, at long last i got them working, and this stuff looks quite nice now 👌2 -
Now this is fucking ridiculous... Our website is being constantly limited though we've never reached even 80% of the available CPU resource.
The hosting said that we had the CPU fault (that fucking cyanide spike on the graph that triggers the limit once) because of huge load on the server. The FUCKING SERVER... Not our virtual environment. And once more because of the RESOURCE MONITORING service caused a server restart. For fucks sake, really???
And apparently it's perfectly normal that all users even ones that run in low resources are being limited to a level that a request takes 30 seconds to complete instead of frickin' 1...
The best they could offer is to move us to a new server, which will arrive in two weeks, if the problem persist. IT'S PERSISTING FOR FUCKING MONTHS YOU MORON. I wonder how much time would have been taken you to realise the server shutdown this week if I hadn't phoned you in 5 minutes. FUCK!
Every shared hosting is that garbage or am I just the choosed one?11 -
I have a really old Toshiba NB200-10P notebook with Intel Atom CPU and 2 gigs of RAM and 32 bit architecture... It was made for Windows XP, and now it barely runs Windows 7... So I'd like to give new life to this old piece of hardware with Linux (for basic tasks like Office, and maybe coding with Notepad++)...
I really like Manjaro, but after they ditched 32 bit support, the Manjaro-32 community project seems a bit unstable...
So, could you devs please recommend some reliable - and somewhat good-and-simple-looking - Linux distros?20 -
Okay then, ex-android user there.
It started with Xperia TX - it was flagship Sony phone back then. It blew my mind when I touched it for the first time. You know, exploring android for the first time in my life was amazing.
It ran just well for about a year. Then it started to fall apart. I need to clarify that I kept it non-rooted, full stock. I'm not into that customization things.
At first, I noticed significant lags. They were everywhere. The longer I used smartphone, the more lags I encountered. I did factory reset, but lags haven't gone anywhere.
Year 2. Front camera stopped working. Battery became unreliable as fuck, going down to 40% and then instantly to zero. What?
Year 3. Camera broke. It refused to start, just giving me "Camera is not available" error.
I tried factory reset again. It helped at first, but month have passed and all that issues came back. And it also became sluggish as fuck.
Got Meizu m3s year ago. The exact same story. Long story short, in one year I got this:
1. Black spots on every picture I take. Much likely a matrix issue.
2. Camera also became slow as fuck, requiring about 10 seconds to even start.
3. Vertical stripes all along the screen. I never dropped my phone, it just appeared once and became brighter and brighter every day I used the phone.
4. Two huge yellow spots on screen. I think it happened because phone's cpu heat up the screen and it broke.
But the most important thing is that fucking lags chased me in every app, they were everywhere. Fucking tiny-ass lags. And they're not going anywhere, they're become more and more significant with time.
Don't say me about oneplus, samsungs and other top android phones. They are conceptually the same, the only different thing is hardware.
That's why I switched. IPhone has its downsides, but it's silky smooth. And my friend's iPhone 4 (not s) feels just as smooth as my brand new se.
I'm not going to jailbreak it. I don't need customizing the hell out of it.
I just needed quick and reliable phone, and SE seems to be exactly what I wanted.
Peace to android folks tho✌️17 -
Here, a full retrospective of my Apple products ownership.
iPhone SE – after Android, I was absolutely amazed by how fast it worked. No UI lags, camera works absolutely instantly no matter the light conditions, all the GPU-heavy games work butter smooth.
After camera and charging port failures on Xperia flagship and CPU literally melting through screen rendering it unusable on Meizu, it was enough to make me interested in Apple products.
When I was using Meizu, I actually got a twitching eye which was triggered by UI lags. After two months of using iPhone, I noticed that something was missing – my eye wasn't twitching anymore.
iPhone actually cured me.
MacBook 12 – a 900 grams laptop with passive-cooled mobile CPU running many Chrome tabs, heavy Webpack HMR build, VSCode and Slack just fine. Yes, you can't play games, but I don't even require it from a laptop this tiny.
Butterfly keyboard that internet hates so much actually increased my typing speed and comfort compared to MX Red mechanical keyboard, and ForceTouch trackpad made me forget about mouse. I learned how to disassemble the Butterfly keyboard if I ever need this but the keyboard never failed.
I use this laptop to this day and it still even smells like the day one, a beautiful smell of a new Apple product.
iPhone X – got it because of the camera, stayed for great battery life and amazing OLED display. I use telephoto lens exclusively and it made me lay off my Canon DSLR with Helios lens which stays on my bookshelf covered in dust to this day.
True black of OLED display which is undistinguishable from the screen bezel is stunning. To this day, battery surely works for one and a half days and I watch youtube really often.
I sometimes struggled to unlock iPhone SE with wet fingers, but with FaceID, as soon as I look at the screen the phone is unlocked. Works perfect every time, never had an issue with this.
Stainless steel body feels premium compared to aluminum. Stereo sound is a major selling point if you're like watching videos and playing games on your phone. Overall amazing product and a huge improvement over SE.
Apple Watch series 4 – really comfortable fit. Nice battery life, once I forgot about it for like ten days during lockdown and it was still working, even though on power reserve mode. Really reliable in terms of battery life and liquid protection. Very satisfying Taptic Engine crown clicks. I run every day and Apple watch always measure my heart rate correctly, and the running app is well designed and a pleasure to use. Overall a nice accessory to have if you use iPhone.
Powerbeats Pro – great sound and battery life. I switched from Shure SE215 which was great, but it had wires. I listen to a lot of music so the sound quality is important for me. When I was choosing earphones I visited a store where you can listen to them all. I listened through earphones like Noble Audio Kaiser Encore and JH Audio Layla, and of course $4000 Laylas sound better than $249 bluetooth earphones, but the difference in sound doesn't justify the difference in price to me.
Powerbeats pro is the Apple H1 chip true wireless earphones with largest driver of them all which makes them sound better than AirPods Pro – it's just physics. Bass in Powerbeats is amazing, which is also true for my Shures, but Powerbeats also win in clarity.
It connects seamlessly to both my MacBook and my iPhone, and everyone in voice chats can hear me really good.
Huge case is a major throwback compared to AirPods, but the battery life of earphones themselves is so great that I just leave the case at home and only carry earphones and it works for me.
Apple Link bracelet in space black – really better than I expected. Intricate detailing, literally the steel that Rolex uses, top-notch finishing and polishing – all that for just 450 dollars. I only used it for several days now, but it already feels like a really satisfying product.
Before all that I was using Linux. It took a year for elementaryos devs to fix wifi for my laptop. Ubuntu looks and feels ugly. Pop OS felt like garbage. Manjaro was also just that – garbage. KDE Plasma – I don't even want to talk about that. A monstrocity where you accidentally click a wrong switch in the settings and your system won't boot up again. Also, PulseAudio. Struggles with proprietary drivers and software updates.
Windows? I serviced a lot of Windows PCs through my career and it never, never worked as intended. I'm no dumbass, I always managed the rights correctly and never installed sketchy apps. My latest ryzen gaming build with a lot of ram also lags somehow even in Windows 10 UI.
Before I switched, I defended Linux.
My life was a lie.
I'm sorry to everyone who I offended based on their opinion on Linux.33 -
Holy fuck AMD really have made a comeback... NVIDIA is admitting they lost (how I interpret this, no flame) and then we have Intel literally running their already fragmented CPU range into a ditch of mismatched specs and just brute forcing higher cord counts on an almost out of date manufacturing process...
And now AMD is also releasing new mobile chipsets with integrated Vega graphics (on the zen based chips) that are already outperforming the closest Intel competitor on thermal and raw performance...
What the fuck!!!
AMD might actually become Intel and NVIDIA rolled into one the way they are going in my eyes...12 -
This is all the DHCP settings I get on this shitty router I got from my ISP. The other settings are more the less the same, which is trash. Don't even get me started on the interface.
The thing doesn't even support *WRT firmwares because it has a Realtek CPU.
Thanks. Guess I'll have to buy a new router now...7 -
Just tried the new PYCHARM update
Honestly speaking I was a bit disappointed with the update as a whole.
Firstly, I it didn't recognize conda as the default development environment and moreover it didn't let me switch from virtualenvs also it took away my in built terminal option and to further aggravate my miseries the performance of the IDE as a whole had fallen quite a bit for instance at one point the ide alone was using around 1.5GiB of RAM and a huge
chunk of my CPU usage -
Company paid 1,500 usd per laptop (dell somethings) for them to just be azure access clients
We can't use outlook or teams. Or even IDEs on them to take advantage of the really good cpu and ram. We have to use the OTHER company device for that stuff
Ohhh and all our azure stuff is getting mac restricted so only the new laptops can access so basically forced to jump back and forth all damn day
Sometimes I feel companies enjoy wasting money3 -
!rant
So I have bought a new laptop and this time instead of straight up booting linux I had an idea of giving micro$oft a try, so I have decided to use only their services for 2 weeks.
To be honest, I really did not expect windows to use do much cpu and hdd during updates and background tasks, but after a day it was ok and windows feels snappier than during my last encounrer (maybe cause the new hw?).
I was even so dedicated that I started to use cortana and I have to tell, that she is dumb as fuck, since she fails to understand even the basic tasks and if u want something advanced, she refers to the next update. But boy, tell her to open Visual Studio and she asks if you want VS Code or Visual Studio, which seems great. But my response was 'Code' then she insisted that I said Coke. Im like OK, Im not native english speaker, lets try Visual Studio Code, where she told me that there is no such thing and Spelling VS - Code ended me in bing search for Unesco :/
I really want to like Cortana, she has nice name, nice history, but she is like that A girl from class, who looks gorgeous, has great voice, but then u reallise that she just eats a book before exam and after that she is that dumb basic hoe.
I also gave a shot to Bing and Edge. Bing is something between Google and DuckDuckGo, since it gives you a liiitle less results from search history, yet if you want to find something in different language its even possible to tell you that what are you trying to find does not exist.
But I have to tell, that I like Edge and I mean it. Like... Its fast and has some good features, like pushing all your open tavs away, so you can open them Later. It also does not have that stupid ass feature that lets you control tab from left to right, not by chronological order, so you wont end up in infinity loop of 2 tabs. And even if people make fun of M$ trying to convince you to use Edge by being too aggresive. God go on edge and try to use some Google Service(You still dont use chrome?!).
I also tried to play with .Net core and I have to tell that against java they are a bit further. I liked some small features, but what I just simply loved was rhe fucking documentation. You basically dont need google, sincw they give you examples and explain in a human way.
What I didnt quite get was the 'big' Visual Studio. Tje dark theme to me feels strange(personal and irrelevant). Why the hell I do need to press 2 shortcuts to duplicate line?! Why is it so hard to find a plugin to give me back my coloured brackets and why the fuck it takes like a second to Cut one line of code on a damn i7?!
Visual studio Code was something different. It shows how dark theme should be done, the plugin market is full of stuff and the damn shortcuts are not made for octopi. So I have to recommend it ^^.
I even gave a shot to word and office as a whole and fuck I never knew that there are so many templates. It really made my life easier, since all you need to do is find the right one in the app, instead of browsing templates online, where half of them are for another version of your text editor.
Android Launcher was fast, had a clever widget of notes and the sync was pretty handy to be honest so I liked that one as well.
What made me furious was using the CLI. Godfucking damn what the fuck is ipconfig?! :/
Last thing what made me superbhappy was using stuff without wine and all of the addional shit. Especially using stuff like Afinity Designer and having good looking apps in general. I mean Open source has great tools l sometimes with better functionality. But I found out, that what is pleasure to look at, is pleasure to work with.
To Summarize a bit.
It wasnt that bad as I expected. I see where they are heading with building yet another ecosystem of It just works and that they are aiming at professionals once again.
So I would rate it 6/10, would be 7 if that shit was Posix compatible.
I know that for Balmer is a special place in hell... But with that new CEO, Microsoft at the end may make it to purgatory..5 -
I just spent an hour and a half installing a cpu cooler for my new server. I got so pissed my brain just switched to happy mode and I genuinely felt good and started laughing like a psychopath for no reason.
All in all, terrible experience, fuck cpu coolers. I've had terrible experience with them throughout the years.5 -
I have had to work on a project with a pc104 stack running yocto. I have had this since December last year and the image has always just randomly crashed 🤔. Yesterday I found out why!!
I am able to read the sensor of the cpu temp this has never been over 60/70 degrees C (yes I am English), however after running multiple tests and finally hitting my last wits I made the Kernel output over serial as no msg was shown on crash.
The company we have got the HW from always said this board won't over heat it throttles the cpu blag bla bla... And you guessed it in the mid of nothing but mess was a message "thermal_zone0 critical temp 127 degrees shutting down"
I didn't know if I was happy or about to cry as I didn't know if after working on this project for the last 6months I was back to the drawing boards as I need new HW or my gut at the start of not trusting the Company we are using!
Needless to say I have no idea what Monday will bring, I will keep you all posted as we all do care!
Much love
Jim -
!Rant
Hello community!
I have a some Questions for you
I want to "Convert" my Prebuild PC into a Normal case...
My currency build has
A GTX 1060
An M.2 SSD (Idk the Brand)
A 2TB Hard Drive
A Intel Core I7 6700 (non K)
16 GB SO-DIMM DDR4 RAM.
And I Just want to put everything into an standard case (Idk which one Right now)
And also want to Upgrade my CPU to a I7 7700K
So I need,
A MoBo,
A PSU,
Case,
Cooler,
And new RAM
(Budget around 500-800CHF)
(Its low but I think its manageable)
Thank you for your suggestions in your comments!9 -
!rant
Got my new pc up and running, asrock x370 killer sli/ac mobo, ryzen 2200g cpu and radeon rx550 with 12g ram (don't ask). Nothing fancy but sure beats my old am2 system! New pc is so quiet :D9 -
Acer vs MSI Laptops.
Five years ago I bought an acer aspire vn7-591 laptop in Redcoon. It was the expensive laptop I bought ever in those days. My experience at the beginning was really bad because the battery laptop crashed after few months and the screen had some vague/dead pixels, but the worst was the imminent bankruptcy of Redcoon. So I couldn't use the warranty. Anyway, It didn't bother so much I have been enjoying this laptop and still doing it. However, last year the screen put me on alert since it started to fail with vertical bars and color changes.
It was time to buy a new computer and due to the problems with some of the components, I've decided to buy a laptop from a company with a better reputation than Acer in aspects as the reliability of the components.
My choice was a Msi Prestige 15 because of the thunderbolts, since the rest of specs are 'more or less similar', although it has more updated hardware, it is lighter, battery holds up to 4-5 hours etc... But... It is really noisy compared to my Acer computer. 2 CPU fans are around 3000 rpm in idle state... Acer seems to be working without using the fans if you are doing intensive work. I google it as I thought it was a factory problem, but it seems to be not a malfunctioning... In fact I found other users complaining about the same and the community proposed to reduce the fan speed through software.
Right now I have both laptops working and since the new boy is in house, Acer is working flawlessly. I am preparing the Acer for my girlfriend as a gift, otherwise it is a pitty to shut it down and store it in a wardrobe.
So, this is my impression about ACER and MSI. I m still experiencing with the new laptop, but I find weird things like the fan speed or how hot it gets in idle despite it uses a new generation of intel i7 cpu with lower consumption... I should monitor the power consumption...8 -
#need_help
Dear all,
I'm trying to make a choice, a choice that won't make me regret it for the few years advanced, I'm in a dilemma, I don't know which MacBook should I get for my everyday life, I currently work as an iOS developer (Learned iOS using all kinds hackintoshes, yeah I never bought a single apple computer, yet), and always have motivation to learn new stuff (from machine learning, to web development, to making games with unity (or whatever engine), hell I even like to design stuff from time to time using Photoshop, sketch, I sometimes do video editing using premiere and after effects), and I yet have to choose which laptop to get, I got only one week to make the choice so...
Here are the options:
The new MacBook Pro 2016 (Touch Bar edition):
Pros: 'Latest' and 'greatest', have thunderbolt ports which makes it (sort of) future proof, TouchId for unlocking the laptop using a fingerprint.
Cons: You need a damn dongle everywhere, no escape key (Which I use for the autocomplete feature in Xcode), and this touch bar (Which I really have no idea if i will ever use it other than the nyan cat app for 5 minutes), plus I heard about battery issues with it (don't know if they resolved it or not), fucking huge trackpad, and no fucking MagSafe!
The previous model MacBook Pro 2015:
Pros: Ports, lots of them, small trackpad (Which you don't have to worry about your palm screwing up your work), and MagSafe! (Which I honestly don't know if it'll make any difference for my usage)
Cons: has old CPU from Haswell generation (I know that it won't feel different, it's just that I like to have parts that are the 'latest')
Now some questions, for people who have the old MacBooks and new MacBooks:
For the ones with old MacBook:
If you were given the choice to replace the old MacBook for the new one for free, would you go for it?
After all this time, how's the battery performance? is it still great from the time you bought it?
Foe the ones with new MacBook:
Does the huge-ass trackpad interfere your work day?
Do you miss magsafe to a point where you really want to throw out the new laptop and go back to previous model?
Did you get used to carry out dongles everywhere?
Did you like the TouchBar? Does it help you in your everyday work? from designing to coding to whatever, do you think that now you can't live without it?
How's the battery performance?
Is programming on it joyable? or the new keyboard and touchpad are just a meh?
Strawpoll to make it easier to vote:
http://www.strawpoll.me/12856510
In addition to that I would love that you guys detail me your experience and answer some questions that I posted above, I would be very, very grateful.2 -
So I recently finished a rewrite of a website that processes donations for nonprofits. Once it was complete, I would migrate all the data from the old system to the new system. This involved iterating through every transaction in the database and making a cURL request to the new system's API. A rough calculation yielded 16 hours of migration time.
The first hour or two of the migration (where it was creating users) was fine, no issues. But once it got to the transaction part, the API server would start using more and more RAM. Eventually (30 minutes), it would start doing OOMs and the such. For a while, I just assumed the issue was a lack of RAM so I upgraded the server to 16 GB of RAM.
Running the script again, it would approach the 7 GiB mark and be maxing out all 8 CPUs. At this point, I assumed there was a memory leak somewhere and the garbage collector was doing it's best to free up anything it could find. I scanned my code time and time again, but there was no place I was storing any strong references to anything!
At this point, I just sort of gave up. Every 30 minutes, I would restart the server to fix the RAM and CPU issue. And all was fine. But then there was this one time where I tried to kill it, but I go the error: "fork failed: resource temporarily unavailable". Up until this point, I believed this was simply a lack of memory...but none of my SWAP was in use! And I had 4 GiB of cached stuff!
Now this made me really confused. So I did one search on the Internet and apparently this can be caused by many things: a lack of file descriptors or even too many threads. So I did some digging, and apparently my app was using over 31 thousands threads!!!!! WTF!
I did some more digging, and as it turns out, I never called close() on my network objects. Thus leaving ~30 new "worker" threads per iteration of the migration script. Thanks Java, if only finalize() was utilized properly.1 -
People going crazy over the new Intel CPU exploit. I'm just sipping coffee looking at my AMD CPU that's never done me wrong 🙌15
-
So today my company was removing most workspaces with USB 2 connections, DP cables and magsafe 2 power cables. This means that my MBP mid 2014 can't connect to the keyboard and monitors anymore. It already struggled with 4K, so my 2K options were already limited, but now the last few spots are mostly gone. In short: I'm being forced to upgrade.
But tell you what: I don't want to. It feels like a waste to recycle my laptop (even if it's company paid and owned) while it's perfectly acceptably fine. And mind that I will get the latest and greatest i9 for free. Yes, that overheating, throttling failure of hardware design piece of shit. 2 coworkers already own the beast and confirm that it gets really hot really quickly. One of them even has daily crashes (the laptop just turns off) and random reboots. A total waste of money. And my future time. As if it's not enough work to migrate to a new laptop (even with Time Machine).
So, fellow ranters, what do I do? I hope I can leverage the second best MBP (CPU-wise) from this situation, unless there already is a bunch of i9s in the office ready to be used. I really, really don't want one. And I think my current computer is great for what it is, even if it's old. It's a really pro machine for my needs (I'm very efficient, except for Android Studio).
I even consider asking for a Linux machine, but then a whole new world opens to me that may be a step too big (since I barely have hands-down experience).
Enlighten me with your ideas, muggles!5 -
Budgie up and running. Tho, I had an issue overload cpu, and stopped previews-creator. Is that even safe?
Anyways, half of my customized deepin shit doesn't work properly anymore. It's a piss off. Wondering if a downgrade would help.
If shit stops breaking for six more months, I might be able to finish uni and go back to workforce properly, where cash-cash-dollar can buy me a new laptop, or a workstation that allows me to duck around with current laptop.
Like, being a student again has been so unnerving. Won't do it ever again. (a PhD doesn't count as study 😜)2 -
N64 emulation possible on Nintendo hardware? Are you SUUUUUUURE???
According to GBATemp:
Wii:
CPU: Single-core PPC, 729MHz
GPU: 243 MHz single-core
RAM: 88MB
Possible? Yes (always full-speed)
New 3DS:
CPUs: 2, Quad-core ARM11 (804MHz), Single-core ARM9 (134MHz)
GPU: 268MHz single-core
RAM: 256MB
Possible? No
Switch:
CPU: Octa-core (2x Quad-core ARMv8 equivalent), 1.02GHz
GPU: 307.2MHz-768MHz (max. 384MHz when undocked)
RAM: 4GB
Possible? Barely (5FPS or less constant, only half the instruction set possible to implement)
this all makes total fucking sense7 -
I am thinking about building new PC, so I made a list of components. What do you think? Any suggestions?
I don't know which GPU model/brand to get yet. Would like to play at 1440p >60Hz so GTX1070 is not enough if I don't wanna buy new GPU in a few years (probably 4-5 years as I always do).
- CPU: AMD RYZEN 7 2700X
- GPU: GTX 1080 Ti / GTX 1080
- HDD: WD Gold 2TB
- SSD: Samsung 860 EVO 250 GB
- RAM: Patriot Viper RGB Series 16 GB KIT DDR4 3200 MHz CL16 DDR4 Black
- MB: ASUS ROG STRIX B450-F GAMING
- CAS: NZXT H500 Black
- PSU: Corsair HX75012 -
Fucking hell my insights are late ones...
So I am working with fluid dynamics simulation. I went home fired up the laptop and started the calculations. This is how the events went:
9 PM: starting the calculation
10 PM: checking on the graphs to see whether everything will be alright if I leave it running. Then went to sleep.
2 AM: Waking up in shock, that I forgot to turn on autosave after every time step. Then reassured myself that this is only a test and I won't need the previous results anyway.
5 AM: waking up, everything seems to be fine. I pause the calculation hibernate the laptop and went to work.
6:40 AM on my way to the front door a stray thought struck into my mind... What if it lost contact with the licence server, while entering hibernated state. Bah never mind... It will establish a new connection when I switch it back on.
6.45 AM Switching on the laptop. Two error messages greet me.
1. Lost contact with license server.
2. Abnormal exit.
Looking on the tray the paused simulation is gone. Since I didn't enabled autosave, I have to start it all over again. Well. Lesson learned I guess. Too bad it cost 8 hours of CPU time.2 -
Programming is like religion.
- you are making sombody to obey your will
- you have limited amount of resources (churches, processors, RAM, holy items)
- when you torture it, usually it will comply but not in way you are expecting
- smashing nail with hammer is not good idea (you can run out of resources again. Getting new CPU is costly as same as getting new believer)
So, who's in for new Church of Unhandled Exception?1 -
So I bought a gtx1650 gpu for my old phenom II X4 pc. It didn’t work – the screen vent black in like five minutes after powering up the pc.
I was disappointed, but instead of returning the gpu, I bought all the other components to build a new pc on ryzen cpu. Including the gpu, it all was like $400 and I still have all my old parts to sale.
Now I’m here, playing all the latest games like doom and wolfenstein on ultra in 1080p 60fps and I’m more than happy.
I basically found a way to convert my bad experience into good experience. I’m just off my therapy, so all that bad experiences that may seem insignificant are a big deal for me.
I didn’t knew it was possible to make a good emotions out of bad emotions that easy. If only I knew the way to apply this strategy for any arbitrary situation.
(please miss me with that boomer bullshit like “nothing is wrong stop whining and get over it” etc. I’ve been there, I’ve done that and I needed medical treatment afterwards. “Getting over it” just doesn’t work)6 -
Fucking shit! So I built this new gaming rig: https://devrant.com/rants/1795588/...
....and fuck!
Firstly the RAM does not fucking run on 3200MHz. The maximum stable speed is 3000MHz...
A secondly, the CPU is so fucking HOT! 50 degrees Celsius in BIOS, underclocking when I try to run stress test. I knew it is thermal paste, so I decided to take the cooler off and see and buy a new better paste and wtf AMD, their paste is so shit, that there was actually no layer of paste on the CPU, only on the edges big piles of it. WUT?4 -
Can anyone help me with this theory about microprocessor, cpu and computers in general?
( I used to love programming when during school days when it was just basic searching/sorting and oop. Even in college , when it advanced to language details , compilers and data structures, i was fine. But subjects like coa and microprocessors, which kind of explains the working of hardware behind the brain that is a computer is so difficult to understand for me 😭😭😭)
How a computer works? All i knew was that when a bulb gets connected to a battery via wires, some metal inside it starts glowing and we see light. No magics involved till now.
Then came the von Neumann architecture which says a computer consists of 4 things : i/o devices, system bus ,memory and cpu. I/0 and memory interact with system bus, which is controlled by cpu . Thus cpu controls everything and that's how computer works.
Wait, what?
Let's take an easy example of calc. i pressed 1+2= on keyboard, it showed me '1+2=' and then '3'. How the hell that hapenned ?
Then some video told me this : every key in your keyboard is connected to a multiplexer which gives a special "code" to the processer regarding the key press.
The "control unit" of cpu commands the ram to store every character until '=' is pressed (which is a kind of interrupt telling the cpu to start processing) . RAM is simply a bunch of storage circuits (which can store some 1s) along with another bunch of circuits which can retrieve these data.
Up till now, the control unit knows that memory has (for eg):
Value 1 stored as 0001 at some address 34A
Value + stored as 11001101 at some address 34B
Value 2 stored as 0010 at some Address 23B
On recieving code for '=' press, the "control unit" commands the "alu" unit of cpu to fectch data from memory , understand it and calculate the result(i e the "fetch, decode and execute" cycle)
Alu fetches the "codes" from the memory, which translates to ADD 34A,23B i.e add the data stored at addresses 34a , 23b. The alu retrieves values present at given addresses, passes them through its adder circuit and puts the result at some new address 21H.
The control unit then fetches this result from new address and via, system busses, sends this new value to display's memory loaded at some memory port 4044.
The display picks it up and instantly shows it.
My problems:
1. Is this all correct? Does this only happens?
2. Please expand this more.
How is this system bus, alu, cpu , working?
What are the registers, accumulators , flip flops in the memory?
What are the machine cycles?
What are instructions cycles , opcodes, instruction codes ?
Where does assembly language comes in?
How does cpu manipulates memory?
This data bus , control bus, what are they?
I have come across so many weird words i dont understand dma, interrupts , memory mapped i/o devices, etc. Somebody please explain.
Ps : am learning about the fucking 8085 microprocessor in class and i can't even relate to basic computer architecture. I had flunked the coa paper which i now realise why, coz its so confusing. :'''(14 -
I remember that when I was about 8 years old, my dad brought a desktop computer home one day.
I don’t remember any specs, but it had a huge ass CRT monitor, a very loud clicky keyboard, a mouse with a real ball inside, and a CPU that uses floppy discs and CDs. Nope, CDs, not DVDs. And on that computer, it ran Windows 95. There’s was no internet most of the time (it was still quite expensive and unnecessary and dial up was troublesome to set up back then).
I remembered playing bootlegged games sold in CDs that my dad bought during his trips to China back then. Duke nukem, Command and Conquer Red Alert 2, Microsoft solitaire and GTA 3. Those were the games I played.
As a kid, it was glorious, looking through a box on a table, seeing and interacting with so many different worlds, stories, characters and games. I really miss those simpler times.
These days, every time I open my laptop, and I see that new mail that need to be dealt with, that homework that’s about to be due and a reminder of my next class in 15 min. Well shit.1 -
TL;DR Does MacBooks degrade faster for developers due to poor thermals?
I’m developing on a 15” MacBook Pro for work. I got it new last year. Now I’m experiencing that it crashes when punishing my CPU with my hardly CPU optimized scripts.
My thought that the poor thermals MacBooks has could be the reason. I mean, Macs are sort of known for their reliability, however I punish my CPU a lot more for many more hours, every single day than the average MacBook user.
Could the instability really be due to a fact that last gen 15” MacBook Pros have poor thermals, thus bad design for programmers, making the CPU unstable due to degradation?9 -
What are people unfiltered thoughts on apples new ARM processor? Especially if you are a Mac user?
(I'm an OS hopper, though my main machine is running ubuntu right now, and work machines are usually mac with windows with wsl)20 -
Had a dream about computers on earth mostly stopping working for no apparent reason, yes, again. But this time, they still work on Mars, so we go there, at least some of us. UAC-esque, Doom 3-ish aesthetics, but in a good way, no death and no darkness. No hell plot though, we’re all fine. Both earth and mars are equally semi-livable, but in different ways. For some reason, we can’t ship new CPUs to mars, and 775 pentium considered a good CPU. We use SQL and HDDs. Elon is also there, but he’s nothing, a peasant compared to other scientists and engineers who are a part of the exodus. I had some problems with food and shelter initially, but @netikras helped me2
-
So I wonder if anybody on here has ever come across the LG G3 flickering screen issue.
I bought a LG G3 about 2 years ago. Fantastic phone.. Until all the issues started appearing.
First of all, the glass started lifting off on the left side of the phone, exposing the backlight layer, because the phone somehow managed to bend (insert iphone joke) inside my pocket. I was okay with it since it was a minor visual thing and didn't affect anything.
About 6 months ago the phone started lagging like crazy, and it kept getting worse and worse. It's so bad right now that Twitter will rarely fullscreen images, apps crash all the time, and occasionally the phone freezes to the point where it won't even react to the lock button.
About 2 weeks ago I started getting messages that my sim card was removed, and the phone starts rebooting itself. This would sometimes happen 10 times a day.
I was already pissed at all of these issues and in a desperate need of a new phone, but on Sunday a brand new issue appeared!
The phone's display randomly shut off, then wouldn't want to turn on, and occasionally turn on just to fade out or flicker away.
I of course went to google to see if anybody else has this issue... And it turns out that youtube is full of videos about it. Apparently the overheating issue these phones have slowly ruins the solder on the chips, which in turn creates a poor connection and causes these issues. The only way to solve it seems to be to reflow the chip, but others that tested that only got it to work for around 2 weeks before the issues started reappearing. I haven't tested it myself yet, but after disassembling, fucking around with it, and putting it all back together 3 times I sort of got it to work to the point where I can use it for several minuted before the displays fails. While writing this rant the display only quit on me once at the very beginning.
What I'm wondering is, why is it that nobody at LG decided to address the heating issue by perhaps throttling the CPU more? And I heard the G3 isn't the only LG phone with these issues?
This crap made me lose all my trust in the company. I wanted to upgrade to a G6 or V30, but because of this crap I think my next phone might be a Xiaomi.
This rant is now so long that it's barely even a rant or on topic. I think I should end it here since I have nothing more to say other than the LG G3 is a beautiful but crap phone, oh, and the new iPhone 8 is a flop9 -
*rewrites rust mpsc*
you did it wroooong
I thought my threads were locking if I had thousands of jobs spawning thousands of more jobs. turns out it's fine. actually if I organize my data locks in the way everyone wants to do them my CPU fans go off but my original way you don't feel jack shit and processes faster
turns out it's because 320k jobs is a bit much for mpsc. because my jobs can spawn more jobs the whole thing just grinds to a halt. and there's sync-mpsc which allows you to have a maximum number of data you send through it, therefore I can just have 245 sent jobs instead of 320k but then this locks all the threads because for a thread to finish it needs to finish sending jobs, but a sync mpsc won't let you send a job if current jobs are over the specified limit. so all the threads get stuck sending jobs. smart. not. what's even the point of that?!
and evidently there's no built-in way to prioritize certain jobs. the AI thinks you should just send jobs in and each thread should have a priority queue. I don't know sounds dumb to me. then you could by random luck have threads with lots of jobs that need to be prioritized to be done and other threads stuck hanging waiting for previous jobs / the other threads. no thanks
so clearly the solution is to rewrite mpsc but allow prioritization when a thread goes in to ask for a job to do
since my jobs are intended to start other jobs, it makes sense to have no actual upper bound limit to the number of jobs in the queue but to favour doing jobs that won't start new jobs to lower the RAM and compute necessary to juggle all this
hope this is the actual problem. cuz the code works for like 200 jobs spawning 500 jobs each, which is 100k jobs total
but it stalls to a halt doing 8300 jobs spawning 500 jobs each (which if I do the math -- in my tests it stalls at 320k jobs and seems the number should be 4,150,00 jobs -- yeah I think this is probably the damned problem)8 -
Sigh...this is kinda stupid.
I'm getting a new ThinkPad at work after 4 years. At first I was like "oh yeah...a new machine!". But they are replacing my quad core T540p with a dual core T560. The T560 CPU has a 30-40% less multi core benchmark score (surprise).
So...dear IT: We are not a small 50ppl company that builds some console apps or small shiny hipster web sites. We are developing fucking large business applications with dozens of projects. Our IDEs and our compiler platform are benefiting from raw CPU power and multiple cores. So can I pls not getting A FUCKING DOWNGRADE AFTER >4 YEARS FFS? THANK YOU!
(before anyone asks: keeping the current notebook is not an option because of warranty/support contracts)5 -
Finally got around to some real video encoding work on my new computer but noticing it's not blazing fast...
And more work is still handled by the CPU... But I thought video processing is handled by the GPU, which seems to be barely used at all. I'm using Handbrake but I thought the whole point of dedicated GPU was for intensive graphics and video processing?6 -
!rant/story:
Aaayoo issya boi the OG rapper straight outta Compton, wassup?!
Nah, for real, though.: How are y'all doing? It has been a long time. I hope y'all are having a great time (can sense some peoples' incoming negative comments due to corona).
I built my completely new first gaming rig like a few weeks ago after my school laptop stopped showing any signals.
The specifications are the following.:
- Mobo: MSI B450 TOMAHAWK MAX
- CPU: AMD Ryzen 5 3600 (no iGPU)
- dGPU: KFA2 RTX 2070 SUPER OC
- 2nd dGPU: GTX 550 TI (this one gets a new rant/post)
- RAM: 2x G.SKILL Ripjaws V 8 GB 3600 MHz
- PSU: Be quiet 650 W (It was the platinum edition afaik. I originally bought the 600 Watt Gold edition, but somehow they sent me the 650 Watt for no additional charge... which is great for me haha)
- HDD: 2x 4TB NAS HDD in RAID1 configuration (to load all of my games around 2 TB right now. Got the two HDDs for 110 Euro including the SATA Data cables)
- SSD: One Intenso 240 GB SSD and a reused Samsung 256 GB SSD from my old broken Lenovo laptop
- Case: Be quiet Pure Base 500 black with a glass panel on the side
While building my first PC, it was one hell of a challenge. I knew how it all was working in theory, but to put it all together, practically, was a bit of a challenge, but it was a nice challenge. I learned a lot and have a performance gain I could only ever dream of.
I used to play my games first on a Toshiba satellite l750d laptop (around 40 fps in cs:go with the lowest gfx settings) and then on a Lenovo e51-80 laptop (around 30 fps max minecraft with no shaders and texture packs).
Now I play cs:go on Kubuntu with 400 fps at peak with ultra high graphics. It is unbelievable.
I couldn't trust the system when I turned on the fps display the first time I saw it.6 -
Fucked up my old phone (picked up an overheating phone with super cold hands and the screen exploded) and my new one is way faster despite having almost exactly 1/2 the CPU power (8-core 2.13GHz in old, 4-core 1.4GHz in this one) due entirely to the Android build not being dogshit. They both only have 2GB of RAM, so i'm still sitting at like 300MB of free RAM with nothing running, but it's a start.7
-
I hate Intellij Idea but it's best option available to develop in Scala. Improvements in VSCode/Metals is my last hope.
The (few) things I NEED from Intellij:
* Very good autocompletion
* Refactoring tools (renaming, auto imports)
* Search tools (find usages, sub/super-types)
The (many) things I hate of Intellij:
* Layout with panel sizes doesn't behave properly and it scales instead of remaining fixed.
* Tedious 2-hands shortcuts makes the right hand to move a lot from the mouse
* Delays and lag in the UI, freezes on garbage collection
* High memory consumption, high CPU usage and generally slow and cumbersome
* The delay in the UI between commands is so that it's accidentally possible to introduce typos
* Can't move tabs around and organize them as I like
* Ugly font rendering and missing typography settings
* Multi-caret implementation as a different editing mode is annoying because requires frequent switching
* Unnatural code folding regions, why method arguments are not folded with the method?
* Unhelpful support forum, sometimes dismissive answers
* Highlighting of current word under the caret doesn't work
* Very slow editor, can't keep spacebar pressed to move text or it hangs!
* Several settings reset at every update. Like the auto fetch of git
* New features are added and enabled by default which is very invasive
* Some of the features mentioned above are really annoying and it's not possible/not trivial to disable them
* It uses its own compile and several times it highlights false positives7 -
I bought the new dell xps15 57 days ago and now it’s ducked (pun intended).
Last week the screen stopped working. I powered off and back on. Then I get a cpu failure light sequence.
I call dell. To my surprise they have given me next day support for free. The guy comes the next day.
He says he will come between 4-6pm. at 615pm he phones me and says he will be late. I hang out at work to wait for him.
Finally at 730pm he comes and doesn’t have a screwdriver for the laptop. So he leaves to go buy one. 8pm he comes back. It takes him an hour to replace the motherboard by which time I just want to check it works and then go home. It seems good and we both leave the dark office at 930pm.
The next day I notice the sound isn’t working. He also hasn’t closed the laptop properly and there is a dent on the right hand side.
Despite dell giving me next day support it takes a week for them to come back with a solution.
I now have to send it off to them and I’ll be a week without the laptop...
It was incredible when it was working. But laptops aren’t great when they don’t work!
Perhaps I should have got a Mac...4 -
TL;DR; I need your advice regd. a new workhorse of a laptop and ARM/MS Surface10/Laptop6 for this purpose
So my hi-end dell XPS (9350) keeps annoying me with its screen flickering. And it's an 8 year old ultrabook with 16G of RAM that I'm using extensively for development, devops, researching and whatnot. 16GB RAM is also becoming...not enough for all of it.
So I'm passively looking for an upgrade. I like the 13" profile (ultrabook style) and battery life, so I'd like to stay away from gaming laptops.
There have been talks about ARM being the new thing. I always saw ARM as a consumer-grade CPU arch (browsing, movies, music, docs, etc.), but the internet says that the new MS Surface devices will have ARM/Qualcomm built in and can compete with MB Pro in terms of performance (ref.: https://windowscentral.com/hardware...) and they are allegedly released this spring.
I'm not much of a hardware person, I prefer staying on the logical level of things, so I want to ask you, people smarter than me, what do you think? Is it a feasible upgrade for an XPS13 (i7 Skylake/16G RAM/4k touch)? I'll be running code and image builds A LOT, using JetBrains IDEs and doing similar resource-intensive tasks. I don't care at all about GPUs - I don't use them (integrated graphics has always been sufficient).
What else should I consider?
Any alternatives?
P.S. while I can't stand Windows, I actually like MS's hardware. They are good at making it.14 -
I'm looking into buying a new laptop. I want something speak and ultrabookish, but dedicated graphics are a must.
Will be mostly for work, so a good CPU and RAM are important, but probably will do some gaming. Battery life and cooling are pretty big too. I've looked at the Aero 15X, the XPS 15, Razer Blade, etc. I just can't decide. Need to get it this year for the the tax deductions for my contracting income. Any thoughts? What's the best one out there?12 -
Today at work, I had to do a dry run of our new production environment config for our VM server.
So I setup ESXi in a vm in my workstation. Installed it it, no problem. Setup some VMs, still going good. Then my supervisor came and said I have to make some tweaks and use vCenter. I thought: meh, that came late but ok. Download the vCenter installer from our internal CDN and ran it. Sorry general config, blah blah blah. "Setup with embedded or external platform service controller?" -"Embedded" - "minimal requirements: 11GB ram, 250GB disk storage, 2 CPU cores". Well... Fuck me. My workstation specs: 8 GB RAM, 128 GB SSD, 4 cores.
What the bloody fucking hell does this stuff need 250 GB disk space for?!
Well at least I've got a permanent upgrade for my workstation.1 -
Just noticed that my watercooling Pump stopped working a month or so ago...
It seems that after i installed my Soundcard the molox Connection was broken or didn't get enough juice anymore....
Connnected a new molex cable to the powersupply and everything is fine again.😌
How the f did my ryzen CPU survive this?!?
Just noticed because rising noise from the fans and Bubbles in the CPU Block.😑
Holy f that was close....😳1 -
So, I was building a new pc for gaming and dev, but I can't choose which cpu to put in.
So I thought let's put it here: i7-7700k or Ryzen 1700X?
Your thoughts?3 -
I work in a repair shop, I have 300+ sticks of RAM laying around, from DDR to DDR3 (DDR3 of 4+ GB and DDR4 are either good for replacement on new computers or are sold immediately so don't count them as laying).
Also I have 200+ hard drives, IDE, SATA, SAS, 2.5, 3.5, 20 GB to 1 TB. Lots of them.
What can I do with all this crap? Not enough matching CPU and MB to make a working system (unless I make a very obsolete one but there's no point in doing so). I also am using single hard drives for backups (only 500+ GB SATA-3 though) so I'm mostly concerned with all that old ram. What can I do with them??8 -
So, the story starts with me getting a job. Full-time job for the first time in my 21 years old life. After short conversation about how amazing this company is, after countless lies and stood questions they decided to hire me. I had to get come on Monday a week later with everything prepared.
So of course I did that and got to my workplace on designated time. Turned out nobody was expecting me, nothing was prepared for a new programmer and everyone seemed angry at me for no apparent reason.
After long talk with my new boss I got some less than 100$ pc with CPU that couldn't handle virtualization and expected me to work on software that needed extensive use of virtual machine.
PC is of course filled with all kinds of spying software that uses most of the resources. IT teams only job is to check if programmers are working their assess off for at least 8 hours a day.
I've filled a ticket about granting me access to Debian machine on the mainframe so I could work. No response for two weeks. I've lost hope already.
I have to work on open space with more than 30 engineers. Screams, phone calls, alarms, all at once, all the time. My colleagues seem to not care and I can't understand how.
I was tasked with rewriting major application because old developer did some half assed piece of burning shit. It took him more than one year, I'm finishing it in less than two weeks.
Of course nobody except for me is preparing any kinds of documentation. I had to reverse-engineer whole API for alarm system.
Salary is less than a junior programmer should earn.
But I'm stuck here for at least a year because nobody's here wants a guy whose only experience is as a freelancer. -
Tl;Dr:
The new windows subsystem for Linux might severely slow compilation time for me.
Microsoft is releasing a preview of WSL 2 which works fundamentally different to WSL 1, which I currently use.
For those who don't know, WSL (or Windows Subsystem for Linux) used to be a compatibility layer, which "translated" Linux syscalls to Windows syscalls. This enables the execution of Linux applications on Windows. The new WSL (WSL 2) doesn't do any of that, instead, it is a highly optimised Virtual Machine.
So don't get me wrong from a performance point of view there is no Issue, RAM and CPU usage is truly astonishingly small and performance of Linux applications is much improved over WSL 1.
BUT, apparently, accessing files stored on Windows through Linux is now piss slow.
Great, truly outstanding.
Why is this a problem? Well, I use WSL to develop c++ Linux applications using CLion, the way this works is that you set up an ssh server in WSL, which CLion uses to do compilations.
One _needs_ to have the project files stored on Windows as otherwise CLion on Windows can't access them.
If I wanted a Linux VM I would have installed one.
Urgh.13 -
I’m working on a new app I’m pretty excited about.
I’m taking a slightly novel (maybe 🥲) approach to an offline password manager. I’m not saying that online password managers are unreliable, I’m just saying the idea of giving a corporation all of my passwords gives me goosebumps.
Originally, I was going to make a simple “file encrypted via password” sort of thing just to get the job done. But I’ve decided to put some elbow grease into it, actually.
The elephant in the room is what happens if you forget your password? If you use the password as the encryption key, you’re boned. Nothing you can do except set up a brute-forcer and hope your CPU is stronger than your password was.
Not to mention, if you want to change your password, the entire data file will need to be re-encrypted. Not a bad thing in reality, but definitely kinda annoying.
So actually, I came up with a design that allows you to use security questions in addition to a password.
But as I was trying to come up with “good” security questions, I realized there is virtually no such thing. 99% of security question answers are one or two words long and come from data sets that have relatively small pools of answers. The name of your first crush? That’s easy, just try every common name in your country. Same thing with pet names. Ice cream flavors. Favorite fruits. Childhood cartoons. These all have data sets in the thousands at most. An old XP machine could run through all the permutations over lunch.
So instead I’ve come up with these ideas. In order from least good to most good:
1) [thinking to remove this] You can remove the question from the security question. It’s your responsibility to remember it and it displays only as “Question #1”. Maybe you can write it down or something.
2) there are 5 questions and you need to get 4 of them right. This does increase the possible permutations, but still does little against questions with simple answers. Plus, it could almost be easier to remember your password at this point.
All this made me think “why try to fix a broken system when you can improve a working system”
So instead,
3) I’ve branded my passwords as “passphrases” instead. This is because instead of a single, short, complex word, my program encourages entire sentences. Since the ability to brute force a password decreases exponentially as length increases, and it is easier to remember a phrase rather than a complicated amalgamation or letters number and symbols, a passphrase should be preferred. Sprinkling in the occasional symbol to prevent dictionary attacks will make them totally uncrackable.
In addition? You can have an unlimited number of passphrases. Forgot one? No biggie. Use your backup passphrases, then remind yourself what your original passphrase was after you log in.
All this accomplished on a system that runs entirely locally is, in my opinion, interesting. Probably it has been done before, and almost certainly it has been done better than what I will be able to make, but I’m happy I was able to think up a design I am proud of.8 -
They dont mention it in any reviews but dell has cooling problems. Yesterday I bought a brand new g5 from a shop and fans were not spinning until 80c from which it started spinning at 4900rpm like a vacuum. Note that I have latest bios so its not temperature table problem in bios.
I found out that there is a driver for thermal power which you can install and for most people it does the work (runs 2200 rpm silent cooling for most of time and 4900 rpm when cpu gpu temp goes above around 60c which is annoying)
So I had to spend entire day on figuring out how to control cooling fans myself. And even then dell limited available speeds in bios at 0, 2200 and 4900 rpm.
Anyways now my cooling fans run at 2200 rpm until 70c and 4900 after. So i get some nice silent performance for 90% of what I do and as for gaming, full fans after 75c is fine. I ran DOOM from steam and max temps Ive got were around 85-90c which is pretty high, so Im thinking of doing a repaste event though im afraid of voiding the warranty.3 -
Had to send my new, shit Nokia phone in for repair as the forced update fucked half the built-in apps. My dad tells me "use my old phone for now, it's terrible." He hands me an LG V20.
Spec relations:
Nokia 3.1 A vs LG V20
RAM: 264MB free w/ nothing running vs 1342MB free w/ nothing running
CPU: 4-core 1.7GHz vs 4-core 2.1GHz
Onboard storage: 16GB/5GB user-accessible vs 64GB/51.5GB user-accessible
Battery life: 7 hours under load to 4 hours under load
I don't think he's getting this back.
(He doesn't use it as "it's so slow" but like... with proper care, the only issue is that switching apps sometimes doesn't fully clean the LCD of the old app and it burns in really fast, but minor annoyances for a WAY better device?)7 -
Microsoft announced a new security feature for the Windows operating system.
According to a report of ZDNet: Named "Hardware-Enforced Stack Protection", which allows applications to use the local CPU hardware to protect their code while running inside the CPU's memory. As the name says, it's primary role is to protect the memory-stack (where an app's code is stored during execution).
"Hardware-Enforced Stack Protection" works by enforcing strict management of the memory stack through the use of a combination between modern CPU hardware and Shadow Stacks (refers to a copies of a program's intended execution).
The new "Hardware-Enforced Stack Protection" feature plans to use the hardware-based security features in modern CPUs to keep a copy of the app's shadow stack (intended code execution flow) in a hardware-secured environment.
Microsoft says that this will prevent malware from hijacking an app's code by exploiting common memory bugs such as stack buffer overflows, dangling pointers, or uninitialized variables which could allow attackers to hijack an app's normal code execution flow. Any modifications that don't match the shadow stacks are ignored, effectively shutting down any exploit attempts.5 -
I think anything from Domas (moveaxeaxeax) is beautiful, but project:rosenbridge is just a new level.
The amount of work, weeks of fuzzing and no documentation to find an entirely secret CPU that acts as a backdoor, and then watching him explain it all. It's purely amazing! -
Fuck sake, had to change the cooling system on my desktop, figured out the new cooling is too big(Noctua) to fit the GPU back, change the HDMI output to the motherboard(my CPU is an APU) and black screen all the time :(
It's fucking normal that as soon as the old GPU is lacking the system is not capable to switch to the embedded one? Fuck me.10 -
Damn computer.
Hey guys!
Have a problem with my custom made desktop.
On games, after 5, 10, 20 minutes the PC freezes and I have to restart.
It's not the graphics card, but I think the CPU is overheating.
It's a new i5 8gen Hexacore, shouldn't be overheating.
Any tools I could use to diagnose what is happening?
Thanks6 -
Are there any cons when choosing a Ryzen cpu for my new pc build, considering programming / development / linux?3
-
*tried to install latest intel GPU driver on a notebook*
Intel: this driver isn't design for your system please contact your device manufacturer
*Check again the compatibility list ... The CPU is present*
Me: okay, let's try this one, using display driver uninstaller to clean the old driver, reboot and launch the new one.
Intel: This program will install the new component graphic driver Intel®
Me: Well ... I guess I win this time ?
Still don't know why the previous one seems to block any installation of I don't manually clean it.
My previous version doesn't seems to be specific to anything ...1 -
Hey guys, ran into some thermal throttling issues with my i7-7700k (100c basically all the time, 800-2000mhz speeds) so I'm looking for a new cpu, anyone down to help?
I'm looking to run 1440p 144hz eventually so I don't mind laying down an average sum of cash for one, just need some suggestions.
Is it worth the switch to AMD? Every forum I look at online has someone buying AMD nowadays
I've tried fixing this i7-7700k by remounting the cooler, reapplying thermal paste, changing BIOS settings, making sure the pump is working, but nothing seems to help.80 -
Im finding parts for my brothers new pc, and i really think that Ryzen 5 2600X is a great choice.
But then one of his friends said that Intel is better for gaming, and suggested an i7-8700k.
The price difference doesnt mean much, my brother just wants whats best for gaming, and now im starting to doubt my own decision.
Which CPU would you think is better for gaming only?
Literally gaming only, not even school . Its only gaming.
Lets not start a war, i just want pros/cons, straight up facts or your experience with either :)12 -
We’re only random people living in random places, speaking random languages, eating random food, sleeping, studying and working random hours. Traveling to random points on a sphere.
Just random range is different.
Just random stuff happens on crossroads of two random dots and the entropy speed ups or slows down.
Nothing special at all.
Just a finite state machine iteration.
I mean the amount of effort we put into explanation of infinity is outstanding.
What if there is no infinity at all ?
What if infinity is just misunderstanding of our interpretation of the world around us. It’s just pixels, resolution, gaussian splatting, quantum state, you name it.
Hey man the world is flat. Just put it to the 2d space. How many space you need from a simulation perspective where your patient eyes can only see up to certain amount of light particles per second on a shitty lens.
Propose a world optimization techniques by slowing down subject perception, tiredness introduced. Compress memory, sleep introduced. Limit neurons, cpu power assigned. Deploy on cloud - put it to life. Exit 0 body failure. Exit 1 suicide. Kill -9 killed by tty from ip EARTH.X.Y
What you can do to make the world around this planet alive? Make it blink.
We developers are lazy and I believe that nature is even more lazy than us.
You think you’re going to elevator right now ? You’re going to the preloader. Looking at the window equals playing video from playback. Never goes live, just precomputed fsm. Cars, trains, airplains ? Preloaders everywhere. Highways to split traffic to cities and communication. The road and cities planning department is a matrix maintenance department. And don’t get me started about space.
Space is empty because it’s not even finished. So they put it all behind glass called milky way. You know how glass looked 500 years ago ? It was milky so it’s milky way so we don’t see shit.
If the space would be finished I’ll be starting writing this text from mars, finished it and sent from earth but no it’s light years guys, light years is not a second for a matter. Light year is a second of the the injected thoughts exchange only. Thoughts of the global computer called generative AI that they introduced on local computing devices called cloud.
Even the preloader system is not present, they left us with the one map and overpopulated demo. What a shit hole.I bet they’re increasing temperature right now to erase this alpha build and cash out. Obviously so many bugs here that his one can’t be fixed anymore. To many viruses.
Hope for 0days to start happening so we can escape using time travel or something.
I bet they cut a budget or something, moved the team to other projects. Or even worse solar system team got layoff off because we are just neurons that ordered to do it. And now we’re stuck in some maintenance mode, no new physics no new thoughts to pursue, just slow degeneration. I would pay more for the next run and switch to other galaxy far far away where they at lest have more modern light speed technology.
What do you think about it Trinity ? Not even worth wasting your time for that. No white rabbit this time.
I do not recommend this game at this stage of early access.
- only one available map despite promises for expansions over the years no single dlc arrived,
- missing space adventures
- no galaxy travel mode only a teaser trailers of what you can do in other “universes”
- developers don’t respond to complains
- despite diversity of species and buildings at first sight world looks to generic
- instead of new features bots with mind manipulation, AB testing and data harvesting was introduced
- death anti cheat mode installed1 -
> picking parts for best possible piecemeal upgrade/sidegrade price with parts made after 2011
> need CPU, mobo, RAM, case
> Ryzen 5 2600, 16? No, 32GB DDR4 of just whatever speed...
> wait, what do you mean new motherboards are $300? Wait.
> Wait, no, how is the one Ryzen motherboard under $100 a shit B450M on sale? The fuck are these prices?
TIL motherboards are volatile as fuck on price. In fact, mobos are less stable in value than most crypto.
Jesus fuck.
(The one B450M i saw is actually now $140-some and a nice X570 is currently $120. Why does the price of these fuckers change *daily?*)
as for case none under $100 are good so i picked some random mostly-mesh ATX mid case over my preferred ATX full case and i'm just gonna pray i can find some fan config that works because at this point i'm done with this shit fuck this, a bottom-barrel mobo, low-tier cpu and a decent case shouldn't be $400 or more12 -
MACBOOK PRO 14.2" M3 PRO & M3 - NEW!
• M3 Pro - 12-core CPU - 18-core GPU - 18GB RAM - 1TB
SSD - INT - Space Black - 2800€
I have never seen more brutal specifications for a laptop in my life.
This 1 laptop is more expensive than 2 new iphones and 1 whole imac desktop PC. Why would someone need this?2