Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "make it a trend"
-
Back when I was in college I had this CS professor who was by far the worst I can remember. The class was some bullshit 100 level required intro to CS course, and the guy tried to make it as difficult as possible. Beyond that, he was just a bad professor and did stupid things.
One of the most memorable things he did was give homework assignments, and then in order to collect them (it was a lecture class of about 150 people), he would have everyone pass their printed assignments to the right, and these sheets of paper traveled all the way across the lecture hall in every row of seats. It was a complete mess.
As you can probably guess, he frequently misplaced homework assignments, and many were probably lost through this ridiculous method of turning them in. Some people almost failed this ridiculously easy class because he lost their homework assignments. I think he lost like one of mine so it didn't matter much, but some other people in the class almost failed because of this. I think in the end he had to make a lot of exceptions because of this obvious trend.
Beyond that, he was an older guy who had worked for IBM, and he made that known at least once per class, usually more. "IBM this, IBM that!" So fucking annoying.
I'm glad to be long done with college.6 -
Let's get something straight people, the trend to change terms in programming languages for PC approved ones is NOT for "making the workplace a better place".
If you are one of those who say "oh it's just terms, if it makes them feel better why not?", "I don't care so should everybody else", "the outrage proves we need to change the terms!".
No sir, first of all, since when has programming been about ditching standards to make people "feeel" better? Since when has engineering been about that?! We are engineers, we don't change shit and waste effort trying to fix things that are working.
Second, this word cleansing does NOT come from a well intentioned one, it's not about making the workplace a better place, it's not about minorities, it's about sanitizing language from an ideological and political standpoint to please an agenda pushing minority who doesn't give a shit about any real social issues.
They have done it to movies, videogames, news, political speech, magazines, books and now programming. It doesn't stop and they will never be satisfied, it's not about changing the terms, no one gives a shit about the terms, it's about pandering to ideological crybabies who want to control what you say because it "offends" them or some supposedly oppressed group from which we just hear anecdotal evidence.
Personally I wouldn't give a shit if it was for technical reasons, but it's not and I've seen what this shit does to communities I love and I won't stand it happening to the dev community just because some weak ass, no balls coders decided to pander to the retards on the far left to score virtue points instead of standing their ground.
Are you worried about oppressed groups? Donate money to third world children, speak out about women in Siria, travel to actual shitty 3rd world countries so you realize changing words on a GitHub repo on your expensive ass MacBook, sipping your soy based coffee on an office with air conditioning is not making the world a better place you delusional prick.
You want to ignore the facts be my guest, be willfully ignorant, but I will not police myself and my ideas for your ideological beliefs, not in gaming, not here. Fuck off.31 -
Apple has programmed an avatar maker in iMessage that generates emoji that are supposed to resemble the person. However it does not have any setting for facial structure to represent a person’s sex. I’m pretty sure they did it because “gender is a construct”.
When I was growing up I had some issues with gender dysphoria. I am male genetically. I was mistaken as female my whole childhood because I was “pretty”, my best friends were girls, and I liked cooking, drawing, and dancing. Puberty happened and I started to look like a man. I considered transitioning because I felt female but I decided to let my body do what it wants and do the things I like to do without worrying about if they are gendered or not. I am married and male and I like what I like.
This stupid iMessage avatar. I have tried my hardest to make it look like me. I have long hair and keep my facial hair clean. They don’t have a switch to change some facial physiological traits so I have this Memoji that looks like a woman with a slight hormonal imbalance. It makes me feel conflicted like I felt when I was young. I haven’t thought about it in years and now I feel like I have an uncomfortable secret female avatar that i carry around on my phone and I feel like I’m carrying a secret.
A persons genetics result in differences in facial structures. Biological sex is more than the length of hair and whether a person wears makeup. I hate this “sex is a construct” trend. I’m fine living my life, but then companies push this software onto my phone like propaganda. I want it to look as masculine as I look IRL.19 -
An important message:
PrOpErLy managing servers is HARD.
I get pissed off at customers with ZERO server knowledge who think they can manage their VPS. “Just get a control panel and a VPS” from some flashy provider that makes server management look way too easy.. Clicking around in their fancy control panel, until:
- they need help with their *self-managed* VPS;
- their email ends up in spam;
- they suffer from performance issues;
- they need to restore a backup;
- something breaks, because YES, things break
Way too little people are able to answer:
- when and how do you make backups?
- how do you monitor your servers and which services?
- how do you keep track of trend analysis?
Then I come by with necessary software. SNMP for trend analysis, Graphite for infrastructure health, Sensu for monitoring, Kibana, Ansible for configuration management..
Things that servers need but that customers have never even heard of.. because they can do everything in their control panel..
Until they come crying to me because it broke and they don’t even know how to get into SSH.
I think the ones to blame are VPS providers that tell the tale of how easy it is to install a control panel and never look at your server again.
Customers become responsible for something *business-critical*! Yet they don’t know how it works.6 -
Take the know-it-all guy you grew up with, that ruins every relationship he's ever had with friends and family, because he gets angry when folks don't deem him as the authority, even for shit he doesn't have a single clue about doing correctly.
Now make him the manager of a fast-food restaurant - so he can command anyone he pleases, making them do anything he wants them to, because he feels it's fun to experiment with co-workers emotions.
Give him an assistant manager that realizes that the only way they can keep their job is to kiss his ass, blowing him every once in a while for a ten cent raise, while the rest of the employees do nothing but smile, say "yes, sir", and go about their business - eventually shit talking about him at the parties he's not invited to.
Watch him jump on every fashion trend, no matter how much it costs, until he eventually decides that the job he's had for the last decade and his fellow employees are beneath him, without saving any money to pay for the things he needs to survive, or taking the proper time to learn all the things that would have made him successful in the long run.
Even though he was an uptight twat and a half, some folks feel that he never got the chance he deserved, as death comes knocking at an earlier age than many would have expected; creating an empty, irrational, and partial dependency in their lives, caused by problems he never cared to correct for their love and admiration, while others are happy as fuck that he's breathed his last breath.
This is the state of our current industry.
*Drops the mic*1 -
Windows 10 had one groundbreaking UI innovation, but no one adopted it and even Windows 11 discareded this revolutionary idea:
BUTTONS NEXT TO EACH OTHER AND AT THE EDGE OF A BOX DON"T NEED AN ADDITIONAL MARGIN
Windows 10 was the first and last OS where I never accidentally clicked right next to the X on a window, in a passive area that had no other purpose so it might as well have belonged to that motherfucking button.
I passionately hate this trend, adopted nowadays by every OS, that everything needs to be rounded, separated from the things around it, and "allowed to breathe". They don't breathe. They're not alive. They're fucking UI elements and the space between them is unused, lost space.
The only interaction a button has with its surroundings is that it pushes other content away to make room for itself and responds to the cursor. It doesn't wiggle, it doesn't grow and shrink, and it ESPECIALLY doesn't fucking breathe. Please, just let me click the motherfucking button.
Relatedly, do you know of a good, preferably bluish dark GTK theme that provides window decorations that stretch the full height of the titlebar and are laid out next to each other at the very end of the bar without gaps?8 -
"As it turns out, this world isn't all that complicated. It's pretty simple actually. It's all a game, a very simple game. Of course, some will try to make it difficult. But you can handle them, I know it. I know you can!"
There's a lot of truth in that. When you get into the depths of how the world works, things turn out to be pretty simple.
One thing I cannot rationalize though. The human spirit. The desires that it embodies. I've had this question for so long - what makes us humans human?
If for example a future surgeon - able to exchange individual cells between me and you - would do so, at which point do you become me, and the other way around? 50+% exchange? But that'd mean that at least part of me is still "you". In that state, are you truly you?
Not sure what the cellular definition of an individual is, given that we're headed towards a bionic society where synthetic organs will likely become more relevant than the donated parts of me that I've recently applied as a donor for. I wholeheartedly encourage that future, but the philosophical questions that surround it become more relevant.
How about the impact of influencers on the mind? For example, I've seen the term "certified enganeers" become a trend here, which I'm very grateful for. It does raise a question though. If for example I were to die, would the term live on? And if so, is that a part of what makes me "me"? Would a part of me live on in you? Would your spirit be partially me, due to mere influence?
What makes up the human creature anyway? I think of my own body as a mere vessel for my mind, but I can't quite grasp what makes up the mind, and philosophical questions like "if I were to upload my mind to a robot and instruct it to kill me, would that carbon copy become *me*?"
The human nature is such a weird thing.. and technology doesn't make it any simpler. Is it really just a simple game, with simple rules and e.g. a biological program running inside of a biological motor? Or is there more to it?24 -
I do not like the direction laptop vendors are taking.
New laptops tend to feature fewer ports, making the user more dependent on adapters. Similarly to smartphones, this is a detrimental trend initiated by Apple and replicated by the rest of the pack.
As of 2022, many mid-range laptops feature just one USB-A port and one USB-C port, resembling Apple's toxic minimalism. In 2010, mid-class laptops commonly had three or four USB ports. I have even seen an MSi gaming laptop with six USB ports. Now, much of the edges is wasted "clean" space.
Sure, there are USB hubs, but those only work well with low-power devices. When attaching two external hard drives to transfer data between them, they might not be able to spin up due to insufficient power from the USB port or undervoltage caused by the impedance (resistance) of the USB cable between the laptop's USB port and hub. There are USB hubs which can be externally powered, but that means yet another wall adapter one has to carry.
Non-replaceable [shortest-lived component] mean difficult repairs and no more reserve batteries, as well as no extra-sized battery packs. When the battery expires, one might have to waste four hours on a repair shop for a replacement that would have taken a minute on a 2010 laptop.
The SD card slot is being replaced with inferior MicroSD or removed entirely. This is especially bad for photographers and videographers who would frequently plug memory cards into their laptop. SD cards are far more comfortable than MicroSD cards, and no, bulky external adapters that reserve the device's only USB port and protrude can not replace an integrated SD card slot.
Most mid-range laptops in the early 2010s also had a LAN port for immediate interference-free connection. That is now reserved for gaming-class / desknote laptops.
Obviously, components like RAM and storage are far more difficult to upgrade in more modern laptops, or not possible at all if soldered in.
Touch pads increasingly have the buttons underneath the touch surface rather than separate, meaning one has to be careful not to move the mouse while clicking. Otherwise, it could cause an unwanted drag-and-drop gesture. Some touch pads are smart enough to detect when a user intends to click, and lock the movement, but not all. A right-click drag-and-drop gesture might not be possible due to the finger on the button being registered as touch. Clicking with short tapping could be unreliable and sluggish. While one should have external peripherals anyway, one might not always have brought them with. The fallback input device is now even less comfortable.
Some laptop vendors include a sponge sheet that they want users to put between the keyboard and the screen before folding it, "to avoid damaging the screen", even though making it two millimetres thicker could do the same without relying on a sponge sheet. So they want me to carry that bulky thing everywhere around? How about no?
That's the irony. They wanted to make laptops lighter and slimmer, but that made them adapter- and sponge sheet-dependent, defeating the portability purpose.
Sure, the CPU performance has improved. Vendors proudly show off in their advertisements which generation of Intel Core they have this time. As if that is something users especially care about. Hoo-ray, generation 14 is now yet another 5% faster than the previous generation! But what is the benefit of that if I have to rely on annoying adapters to get the same work done that I could formerly do without those adapters?
Microsoft has also copied Apple in demanding internet connection before Windows 11 will set up. The setup screen says "You will need an Internet connection…" - no, technically I would not. What does technically stand in the way of Windows 11 setting up offline? After all, previous Windows versions like Windows 95 could do so 25 years earlier. But also far more recent versions. Thankfully, Linux distributions do not do that.
If "new" and "modern" mean more locked-in and less practical and difficult to repair, I would rather have "old" than "new".12 -
Static HTML pages are better than "web apps".
Static HTML pages are more lightweight and destroy "web apps" in performance, and also have superior compatibility. I see pretty much no benefit in a "web app" over a static HTML page. "Web apps" appear like an overhyped trend that is empty inside.
During my web browsing experience, static HTML pages have consistently loaded faster and more reliably, since the browser is immediately served with content useful for consumption, whereas on JavaScript-based web "apps", the useful content comes in **last**, after the browser has worked its way through a pile of script.
For example, an average-sized Wikipedia article (30 KB wikitext) appears on screen in roughly two seconds, since MediaWiki uses static HTML. Everipedia, in comparison, is a ReactJS app. Guess how long that one needs. Upwards of three times as long!
Making a page JavaScript-based also makes it fragile. If an exception occurs in the JavaScript, the user might end up with a blank page or an endless splash screen, whereas static HTML-based pages still show useful content.
The legacy (2014-2020) HTML-based Twitter.com loaded a user profile in under four seconds. The new react-based web app not only takes twice as long, but sometimes fails to load at all, showing the error "Oops something went wrong! But don't fret – it's not your fault." to be displayed. This could not happen on a static HTML page.
The new JavaScript-based "polymer" YouTube front end that is default since August 2017 also loads slower. While the earlier HTML-based one was already playing the video, the new one has just reached its oh-so-fancy skeleton screen.
It would once have been unthinkable to have a website that does not work at all without JavaScript, but now, pretty much all popular social media sites are JavaScript-dependent. The last time one could view Twitter without JavaScript and tweet from devices with non-sophisticated browsers like Nintendo 3DS was December 2020, when they got rid of the lightweight "M2" mobile website.
Sometimes, web developers break a site in older browser versions by using a JavaScript feature that they do not support, or using a dependency (like Plyr.js) that breaks the site. Static HTML is immune against this failure.
Static HTML pages also let users maximize speed and battery life by deactivating JavaScript. This obviously will disable more sophisticated site features, but the core part, the text, is ready for consumption.
Not to mention, single-page sites and fancy animations can be implemented with JavaScript on top of static HTML, as GitHub.com and the 2018 Reddit redesign do, and Twitter's 2014-2020 desktop front end did.
From the beginning, JavaScript was intended as a tool to complement, not to replace HTML and CSS. It appears to me that the sole "benefit" of having a "web app" is that it appears slightly more "modern" and distinguished from classic web sites due to use of splash screens and lack of the browser's loading animation when navigating, while having oh-so-fancy loading animations and skeleton screens inside the website. Sorry, I prefer seeing content quickly over the app-like appearance of fancy loading screens.
Arguably, another supposed benefit of "web apps" is that there is no blank page when navigating between pages, but in pretty much all major browsers of the last five years, the last page observably remains on screen until the next navigated page is rendered sufficiently for viewing. This is also known as "paint holding".
On any site, whenever I am greeted with content, I feel pleased. Whenever I am greeted with a loading animation, splash screen, or skeleton screen, be it ever so fancy (e.g. fading in an out, moving gradient waves), I think "do they really believe they make me like their site more due to their fancy loading screens?! I am not here for the loading screens!".
To make a page dependent on JavaScript and sacrifice lots of performance for a slight visual benefit does not seem worthed it.
Quote:
> "Yeah, but I'm building a webapp, not a website" - I hear this a lot and it isn't an excuse. I challenge you to define the difference between a webapp and a website that isn't just a vague list of best practices that "apps" are for some reason allowed to disregard. Jeremy Keith makes this point brilliantly.
>
> For example, is Wikipedia an app? What about when I edit an article? What about when I search for an article?
>
> Whether you label your web page as a "site", "app", "microsite", whatever, it doesn't make it exempt from accessibility, performance, browser support and so on.
>
> If you need to excuse yourself from progressive enhancement, you need a better excuse.
– Jake Archibald, 20139 -
Want to make someone's life a misery? Here's how.
Don't base your tech stack on any prior knowledge or what's relevant to the problem.
Instead design it around all the latest trends and badges you want to put on your resume because they're frequent key words on job postings.
Once your data goes in, you'll never get it out again. At best you'll be teased with little crumbs of data but never the whole.
I know, here's a genius idea, instead of putting data into a normal data base then using a cache, lets put it all into the cache and by the way it's a volatile cache.
Here's an idea. For something as simple as a single log lets make it use a queue that goes into a queue that goes into another queue that goes into another queue all of which are black boxes. No rhyme of reason, queues are all the rage.
Have you tried: Lets use a new fangled tangle, trust me it's safe, INSERT BIG NAME HERE uses it.
Finally it all gets flushed down into this subterranean cunt of a sewerage system and good luck getting it all out again. It's like hell except it's all shitty instead of all fiery.
All I want is to export one table, a simple log table with a few GB to CSV or heck whatever generic format it supports, that's it.
So I run the export table to file command and off it goes only less than a minute later for timeout commands to start piling up until it aborts. WTF. So then I set the most obvious timeout setting in the client, no change, then another timeout setting on the client, no change, then i try to put it in the client configuration file, no change, then I set the timeout on the export query, no change, then finally I bump the timeouts in the server config, no change, then I find someone has downloaded it from both tucows and apt, but they're using the tucows version so its real config is in /dev/database.xml (don't even ask). I increase that from seconds to a minute, it's still timing out after a minute.
In the end I have to make my own and this involves working out how to parse non-standard binary formatted data structures. It's the umpteenth time I have had to do this.
These aren't some no name solutions and it really terrifies me. All this is doing is taking some access logs, store them in one place then index by timestamp. These things are all meant to be blazing fast but grep is often faster. How the hell is such a trivial thing turned into a series of one nightmare after another? Things that should take a few minutes take days of screwing around. I don't have access logs any more because I can't access them anymore.
The terror of this isn't that it's so awful, it's that all the little kiddies doing all this jazz for the first time and using all these shit wipe buzzword driven approaches have no fucking clue it's not meant to be this difficult. I'm replacing entire tens of thousands to million line enterprise systems with a few hundred lines of code that's faster, more reliable and better in virtually every measurable way time and time again.
This is constant. It's not one offender, it's not one project, it's not one company, it's not one developer, it's the industry standard. It's all over open source software and all over dev shops. Everything is exponentially becoming more bloated and difficult than it needs to be. I'm seeing people pull up a hundred cloud instances for things that'll be happy at home with a few minutes to a week's optimisation efforts. Queries that are N*N and only take a few minutes to turn to LOG(N) but instead people renting out a fucking off huge ass SQL cluster instead that not only costs gobs of money but takes a ton of time maintaining and configuring which isn't going to be done right either.
I think most people are bullshitting when they say they have impostor syndrome but when the trend in technology is to make every fucking little trivial thing a thousand times more complex than it has to be I can see how they'd feel that way. There's so bloody much you need to do that you don't need to do these days that you either can't get anything done right or the smallest thing takes an age.
I have no idea why some people put up with some of these appliances. If you bought a dish washer that made washing dishes even harder than it was before you'd return it to the store.
Every time I see the terms enterprise, fast, big data, scalable, cloud or anything of the like I bang my head on the table. One of these days I'm going to lose my fucking tits.10 -
Pull-to-refresh is useless.
If you are a mobile app developer, please get rid of pull-to-refresh. Your users will thank you.
I have the impression that mobile app developers choose to implement the pull-to-refresh gimmick just in order to make their app comply with a design trend. It seems like a desperate attempt to appear "modern" and "fancy", not because of the actual usefulness of the gesture.
Pull-to-refresh is one of those things that are well-intended but backfire. It appears helpful on first sight, but turns out to be a burden.
It takes effort and cognitive strain to avoid triggering a pull-to-refresh. The user can't use the app relaxed but has to walk on eggshells.
Every unwanted refresh wastes battery power, mobile data (if it is an Internet-connected app), and can lead to the loss of form data.
To avoid pull-to-refresh, the user has to resort to finger gymnastics like a shorter swipe for scrolling up or swiping slightly up before down. Pull-to-refresh could even be triggered while pinch-zooming in or out near the top of a page, if the touchscreen does not recognize one of the two fingers.
Pull-to-refresh also interferes with the double-tap-swipe zoom gesture. If one of the two taps are not recognized, a swipe-down to zoom in can trigger a pull-to-refresh instead.
To argue "if you don't like pull-to-refresh, just don't use it" is like blaming a person who stepped on a mine, since the person moved and the mine was stationary.
A refresh button can be half a second away in the menu bar, URL bar, or a submenu, where it is unlikely to be pressed accidentally. There is no need for a gesture that does more harm than good.
Using a mobile app with pull-to-refresh feels like having Windows StickyKeys forcibly enabled at all times. The refresh circle animation sticks to the finger.
If the user actually wants to refresh, pull-to-refresh is slower than a refresh button in a menu if the page is not at the top, meaning pull-to-refresh is useless as a shortcut anyway if the page is in any other position than the top.
An alternative to pull-to-refresh is pull-for-details. Samsung did it in some of their apps. Pulling down against the top reveals additional information such as the count and total size of selected items.
If you own a website, add this CSS to make browsing your website on the pre-installed Android web browser not a headache:
html,body { overscroll-behavior: none; }
Why is this necessary? In 2019, Google took the ability to deactivate the pull-to-refresh gesture on their Chrome browser for Android OS away from users. On Chrome for Android, pull-to-refresh can only be disabled on the server side, not the user side. The avalanche of complaints? Neglected.
Good thing several third-party browsers let the user turn off this severe headache.12 -
Finally getting off my proverbial ass and doing something about the lack of games I like. Going to focus on making an engine for the kind of games I want to play.
No, I am not starting from scratch. Going to base my engine on Godot and use it for my own titles. I am not insane. Making it from scratch is too much work these days. But the indies are shifting from Unity to other engines right now. So a lot of wanted attention will be placed on better alternatives. This means more content and plugin choices will be available to Godot devs.
I kept making excuses as to how hard it will be or it will take forever. It only ended up taking me further away from what I wanted. I have my wishlist of features and I will focus on modularizing them so they can be used as needed. If it makes sense I will make these modules available to the community at Godot. This will help get feedback on what can be improved and generalized further. It will also reduce development costs in the long run. I want to take the approach that No Man's Sky has taken for content and generate as much as I can. I am fascinated by generating objects using algorithms. This seems to be a trend in games.
The struggle I have with games: I want to build things like structures in game (aka Minecraft), I want to build characters in game (aka RPGs), I also want to deform terrain (aka organic voxels), and I want a mixed genre (guns and dragons). Nothing like this exists in a form I want to pay for. I also want to be able to mod the game and for other people to be able to mod the game. That really narrows the list of games down to nothing. Sure there are few games that hit these bullet points, but not all in the same game.
I am finding I struggle to be engaged intellectually at work. I do what I have to for a paycheck. I think having a side project will help with this. One that is radically different than what I do at work is going to be helpful. I need to be realistic about expectations. I probably shouldn't expect any real progress for at least 2 to 3 years and probably more likely 5 years. I have some experience with the tool chains from other engines I have worked with. I also want something that I own and is mine. Even if it sucks.32 -
Antergos is going out of the play. And i saw a very click baity article which poised the following statement at the end:
"Is the death of Antergos a major loss? No, not on its own. Despite the developers bragging about over 900,000 downloads (over the last five years) it’s hardly a popular operating system. Still, its demise is a part of an emerging trend where developers don’t have the resources to continue a project. And both the Linux and Open Source communities should be very worried about that. Developing for love or as a hobby simply isn’t sustainable."
Now, this is, at least to me, bullshitty in the sense that the open source community does not really have anything big to worry about. Large pools of companies would make yeary investments in open source codebases due to the ammount of usefulness they present to their companies. More and more great open sourced projects come out every year OUTSIDE the all eating scope of just web development(which to an extend is fine since it brings communities together)
Saying that a hobby isn't sustainable is funny in itself really.
If people don't have the time to support a hobby project because they are moving on to bigger and better things in shit that actually pays then I am glad for them. It tomorrow Arch, Debian, pop os, ubuntu and fucking freebsd goea out then I would have something to bitch about.
Till then, stating that the community haa something to worry about is just bullshit.3 -
Some two years ago I purchased a license for AV solution on Amazon UK which is my default place for shopping.
When attempting to activate the license, I kept getting this annoying error somewhere along the lines of:
'License from another region.'
I contact the support, they did their magic && the license got applied.
Fast forward some two years - the license is about to expire.
The software is actually good, so I make the choice to renew it.
Thing is, I keep getting redirected to the local site w/o an option of choosing the English language.
I edit the site's address to reach an English version of the site.
On UK's site I am unable to choose my country of residence so I can't complete the purchase form.
I try a few other things && finally reach the Global site where I am able to input the correct data for the purchase, but all attempts to finish the payment fail.
Fine... I'll purchase from the local site.
I purchased the license, activated it w/o problems, but when attempting to download the installer it keeps downloading the localized version which doesn't allow me to change the language.
I contact Support.
S: 'The license you've purchased can be used solely w/ localized version of the software. If you want to use the English version, you'll have to get a refund && purchase from our Global site here [link w/o redirection].'
_Fuck_ this trend of automatic redirection to localized sites && forced localized software.
One shouldn't have to go through all these hoops to get the software in the language of their choice, instead of having the localized version shoved down their throat.17 -
Dev Confession:
I wrote a bunch of code today that created more problems than it solved. I did not commit it. I used git stash to hide it. It took me hours to write. I didn't do a test on a small section of code beforehand. Literally hours of wasted man hours.
At least I didn't commit this garbage into the repo. The approach was fine, but the architecture made it a non-solution. Now I need to redesign this code or leave as is. It is production code I cannot just "change" on a whim.
I have officially dubbed this week as confession week. This should be a world wide thing. People should fess up to their terrible deeds. Lets start a trend and confess to our misdeeds in code and life. Make the world a better place!
What do you say?6 -
!devrant.
More like 'relationship' rant.
The following is gonna be kinda crude and not work appropriate just fyi.
First, whats with this trend of 'shave your pussy", and small tits? Shit just a few years ago it was the opposite! Who wants to fuck a women who looks like an undeveloped boy. Give me a good medium pair of knockers and a (slightly) hairy pussy anyday and I'm in heaven.
On that note, all these guys whining and fucking whinging about "fat" women. Like I'm not chasing deathfat women or anything, but come on dudes. If you're gonna demand 'thin beautiful women' put in the fucking work! Just saying it now. Kinda pudgy, not a deal breaker for me. She have a pretty face, looks good in make up? Fuck it, I'll give her the time of day. Don't come at me with "hurrr...fat is unhealthy! Wheres your standards?"
Wheres YOUR fucking standards? Shit, I could take a 'heavy' girl and in a few months time have her trim. Its about the WORK you put in. All these fucking guys, all these fucking guys man, they all want something for nothing and chicks for free. No one wants to put in the fucking EFFORT anymore. I can't name on one hand the number of 'fat ugly chicks' everyone ignored from school (well except for me), who are now bombshells. If I stayed connected with people maybe I'd be with one of em right now, but I just get tired of this attitude that no one invests any time in others any more. It's all about 'me me me'.
Shes fat, maybe only a few extra pounds? Fat and 'not beautiful'? Fucking DO THE WORK and make her beautiful you bitch! Be her coach, like those fucking instagram couples you see sweating it out. Make her sexy. Become sexy, together.
Get her a fucking treadmill. Get her two. And jog together. Make her *feel* sexy.
More importantly get to *know* her. Why does she get out of bed in the morning? What drive her NOT to get out of bed? When does she feel lousy? What makes her feel that way? In addition to all the other shit men should know how to do, you should learn to play doctor phil, because every girl needs one. Women bond by talking, men bond by doing things together. Relationships should involve both.
Jesus fucking christ, this is basic bitch advice, and it annoys me I keep on coming across these spergs that don't fucking get it. Women are not cars you can stick your key in and just go 'vroom'.
They require maintenance, same as anything, any 'relationship' (because really, what are relationships in 2020, with instagram and fucking tinder and antisocial distancing?). You're a *team*, and i don't mean that in an inspirational way. You're a literal team. And far too many people prioritize the well being and success and concern of 'I' over' 'us'.
In short, if she ain't coming, you shouldn't be either. And if you expect her to on-the-regular put sharp objects on her fucking nether regions, don't expect any blowjobs unless you're doing the same thing for yourself!
Ideally you should be doing it to each other.
After all, you're partners. You trust her to put sharp objects near your groin, right?
Aren't relationships supposed to be about mutual trust?28 -
!dev
I've seen a growing trend in series.
If an author wants to insert a romantic sub-plot into a very much mature series,
He won't write it in a believable way, fitting the tone of series.
Instead he will insert cheap teen-drama and make the audience cringe to death.
What is it with this shit?3 -
Alright, I'll try writing about my recent experience without getting too emotional.
A few months ago, I started a tech job in London and immigrated here for that job. I was glad this company wanted to sponsor a visa, as that was a requirement for me to live here.
Unfortunately, after only a few months in, I learned that the company I joined wasn't quite as nice as I thought it would be. Bullying seemed to be part of the culture. On occasion, I saw coworkers crying. One of my close coworkers was dangerously close to burnout and then "left with mutual agreement". The environment felt like a high school cafeteria. People were drinking heavily early in the afternoon and people were leaving almost at the speed of a revolving door.
I recognized very early on that this was not a healthy environment for me, but as I just signed a rental agreement for a year, and spent a large amount to move here, I was kind of trapped.
Very early on, I was told that the two people before me in the same role were let go right before their probation ended. That scared me off, for reaching out to management or HR. I didn't have the financial needs to lose my job, and due to visa restrictions, therefore would have to leave the country.
When my probation was about to end, and I learned that my performance was good, I decided to provide feedback to my manager. I only mentioned a few things, but still enough. The manager seemed receptive, but it did not seem like he was actually willing to approach the problem itself.
Sometime later, I spoke to HR, explaining some of the issues, and explained my intent to resign. The rep pretended to care, but it did not seem sincere. At the same time, I reached an agreement with my landlord, so I believed I had enough money to safely move out of the country.
A few days after I resigned, the HR rep told me that I owed the company a large amount of money. A part of it was in the contract, which I accounted for. Another part, she was claiming, but was not properly defined in the contract. It said something, but it was confusing. I got a checked later with a legal advisor, and from what I understood, the company would never be able to make me pay that extra amount. This simply because of the contract being so vague.
I told the rep multiple times in the initial meeting about the flaws in the contract, but she ignored everything I said. I then made a counteroffer trying to get her to back off. She then put that in writing, but manipulated my words and kept out all the arguments I made about contract flaws, and my departure being the company's fault.
I didn't receive a reply to my counteroffer for days. It was stressing me out as this could mean I would run out of money soon. Only a few days passed before I got a medical emergency at work just because of the stress all of this caused me.
I saw a doctor and immediately got 2 weeks of sick leave. When I contacted the company again, I was able to terminate my contract, without returning to the office. However, they still didn't want to waive the extra amount of money.
The HR rep pointed out in written communication to my lawyer, something in the trend of "if something wasn't clear in the contract, he should've just asked for details". In that same correspondence, it also stated that they were offering 'as a favor to me' to reduce the extra amount to only a third of it.
Since I never actually wanted to go to court anyway, I decided to settle with that. Now I'm packing to move out of the country, without a job and soon to be completely broke. If I would've stayed where I were and never moved to London, and never worked a day for the past 7 months, I would've had more money on my savings account than I have at this point in time.
I hope I at least learned something from this. I don't think I will move somewhere with a company-sponsored visa again anywhere soon...
Thanks for listening. Ranting does make you feel better :)3 -
This is gonna be a long post, and inevitably DR will mutilate my line breaks, so bear with me.
Also I cut out a bunch because the length was overlimit, so I'll post the second half later.
I'm annoyed because it appears the current stablediffusion trend has thrown the baby out with the bath water. I'll explain that in a moment.
As you all know I like to make extraordinary claims with little proof, sometimes
for shits and giggles, and sometimes because I'm just delusional apparently.
One of my legit 'claims to fame' is, on the theoretical level, I predicted
most of the developments in AI over the last 10+ years, down to key insights.
I've never had the math background for it, but I understood the ideas I
was working with at a conceptual level. Part of this flowed from powering
through literal (god I hate that word) hundreds of research papers a year, because I'm an obsessive like that. And I had to power through them, because
a lot of the technical low-level details were beyond my reach, but architecturally
I started to see a lot of patterns, and begin to grasp the general thrust
of where research and development *needed* to go.
In any case, I'm looking at stablediffusion and what occurs to me is that we've almost entirely thrown out GANs. As some or most of you may know, a GAN is
where networks compete, one to generate outputs that look real, another
to discern which is real, and by the process of competition, improve the ability
to generate a convincing fake, and to discern one. Imagine a self-sharpening knife and you get the idea.
Well, when we went to the diffusion method, upscaling noise (essentially a form of controlled pareidolia using autoencoders over seq2seq models) we threw out
GANs.
We also threw out online learning. The models only grow on the backend.
This doesn't help anyone but those corporations that have massive funding
to create and train models. They get to decide how the models 'think', what their
biases are, and what topics or subjects they cover. This is no good long run,
but thats more of an ideological argument. Thats not the real problem.
The problem is they've once again gimped the research, chosen a suboptimal
trap for the direction of development.
What interested me early on in the lottery ticket theory was the implications.
The lottery ticket theory says that, part of the reason *some* RANDOM initializations of a network train/predict better than others, is essentially
down to a small pool of subgraphs that happened, by pure luck, to chance on
initialization that just so happened to be the right 'lottery numbers' as it were, for training quickly.
The first implication of this, is that the bigger a network therefore, the greater the chance of these lucky subgraphs occurring. Whether the density grows
faster than the density of the 'unlucky' or average subgraphs, is another matter.
From this though, they realized what they could do was search out these subgraphs, and prune many of the worst or average performing neighbor graphs, without meaningful loss in model performance. Essentially they could *shrink down* things like chatGPT and BERT.
The second implication was more sublte and overlooked, and still is.
The existence of lucky subnetworks might suggest nothing additional--In which case the implication is that *any* subnet could *technically*, by transfer learning, be 'lucky' and train fast or be particularly good for some unknown task.
INSTEAD however, what has happened is we haven't really seen that. What this means is actually pretty startling. It has two possible implications, either of which will have significant outcomes on the research sooner or later:
1. there is an 'island' of network size, beyond what we've currently achieved,
where networks that are currently state of the3 art at some things, rapidly converge to state-of-the-art *generalists* in nearly *all* task, regardless of input. What this would look like at first, is a gradual drop off in gains of the current approach, characterized as a potential new "ai winter", or a "limit to the current approach", which wouldn't actually be the limit, but a saddle point in its utility across domains and its intelligence (for some measure and definition of 'intelligence').4 -
Drupal 8 fractured the community, dead ended projects that had years of being built up and supported, started a downward trend in overall number of websites using Drupal when it was still increasing market share, homogenized Drupal with other less successful frameworks that had already attempted it and failed by using composer to replace drush, twig to replace PHPtemplate, and Symfony to butcher Drupal and hang parts of it on.
The mission statement was to "bring Drupal to the modern era" and "be more enterprise friendly". All I've seen them do is make it worse. I have stopped using Drupal now, I still maintain some Drupal 7 sites but now that they killed the Drupal 7 community it's basically dead. Some small attempt was made to salvage it with Backdrop but it will likely never be as big as Drupal was and is mostly dead itself, for one thing it's not directly compatible with the huge library of modules either.
Another thing I loved killed by those without vision and giving into the "industry standards" that make one question the intellect of everyone who subscribes to them being a good idea. But hey that evil procedural programming that worked so long for so many was finally defeated. It's surely better now right... right?
At least this movement was supported by people that can't even tell the difference between the use cases in real projects between Drupal and Wordpress. Software Development is in such a good place and has no hypocrisy. One would never suggest it has lost sight of its original purpose of solving real world problems with computing and become self absorbed with its own navel gazing.
If still in doubt check attached image, it tells a very clear story about how to ruin the life of a CMS. It honestly feels like a hitjob attempted to sabotage it rather than an earnest attempt to improve something that has been doing well since 2001.8 -
Anyway Github has launched its app.
Positioning as a social network and trend aggregator is becoming increasingly clear.
It doesn't make much difference to me, I almost always prefer the web version and Github's works well.
But do you know what we really need?
- Native dark mode2 -
Oh China, you still continue to amuse me... in that special way where i somehow both expect it and am hilariously, breifly shocked... then it's somber, confirming what we know is real/continues to be a societal and cognitive decline trend with no apparent rock bottom, without all-out demise as a near certainty... nor a hail mary play.
... but hey, what better way to digest the real-time info, indicative of something that should be terrifying, but is all too expected, than this unique type of format?
Seriously though, even if it worked amazingly, why would anyone be using it outside in public? Does it require several hours a day? If not, and it was a worthwhile result for you... wouldn't you just make it part of your morning and/or evening routine...even if it had nothing to do with aesthetics, that cant be sanitary... unless you also carry it in a water-tight container or disinfectant and typically bring/use your toothbrush and toothpaste mid-day or at unusual intervals.
I have sooo many more questions about this... and none are relative to who designed/mass produced this, nor the quality of the silicone. As it was developed/produced by the silicone factory ive done great, professional, no bs, business with for about a decade... which is why i waited years to publicly ridicule this contraption.
Fyi- their primary product lines are things like bongs and dab containers; im on the fence of it that makes this better or worse.
Creepy personal truth... i reeeeally wanna know how much that woman got paid... and do to my skill set (ie. Im near utter certainty that i could find her and ask her... likely easily abd definitely without being caught doing anything suspicious. Pro tip: publicly declaring things like this makes it a bit easier to not end up doing it... obvious premeditation adding significantly more to any sentencing.22 -
It seems to be the new trend : building "boxes" based on raspberry pi, including sensors to mesure any sort of thing, and sending data to a REST API.
Was contacted for a project like this, to make the backend for the project.
I ask to the client the credentials of the dev who will makes the embedded dev, to know the format of data I will receive and send to the "box", the client respond that "I don't need to know that", and, besides, they don't have any dev for this post for now, but I can begin the dev for the backend without that, not knowing data structure, and will receive all of that for half December, for a deadline in early January.
Tell the client that his project will never be done in the deadline, got ejected from the project, client is pretty sure he will find à dev who will do all the work in 2 weeks.
Fuckin' startup culture.1 -
I'm currently having a problems sleeping my inner philosopher just keeps thinking about various things. I wanna try to write some of them down as an simply to see what will happen.
I'll write my opinion down as honest as possible so feel free to disagree, but point out what I should rethink, if you want me to consider it.
To me respect has to be earned. I think especially on the internet many people try to skip this crucial step when they try to get respect. Most often when they want an opinion or their ideals to be respected. Most of the time it doesn't even feel like they want to be respected, but rather accepted.
There's nothing wrong with accepted in my opinion, but there are several approaches to get to this point and I despise some of them.
Earning acceptance by earning respect is one of the right ways to do it. Working hard towards your goals, showing your individual strength, standing behind your ideals. These are things I can respect.
I should also mention that these Ideals should be concrete, based on rational thought and a general good will or you will just twist my words to say that I support e.g. IS, Stalin's politics ect.
On a side node, I think it'd be wrong to disrespect everything Stalin did, since, from an economical point of view, he pushed Russia forward by quite a bit.
Then on the other side I see crybabies. People who want to be accepted, without putting effort in their ideals. Most of the time not even aiming for acceptance through respect, but through pity. Honestly, that's all they're going to get from me.
Pity, for their petty ideals.
Basically all I ever see these people doing is attention whoring and practicing multiple deadly sins at once.
Wrath, jealousy, sloth, pride, greed and optionally also gluttony.
Lust is rather a separate package. When I think about it, I link it mostly to horny teens and "send bob and vegane" type of stuff.
Gluttony being powered by sloth or vice versa, enhancing it.
The clear image I have in mind, while I write about this packages of deadly sins however, is that of a jealous person, complaining / getting angry about something they could change change themselves, but want them to be changed for them. Mostly through social networks such as Facebook, Twitter and whatever the fuck Tumblr is supposed to be.
"I wanna be rich, why is <person> richt but I'm not? This world is so unfair 😡". Have you tried working towards becoming rich?
"I don't don't feel pretty. Accept me". Accept yourself. Done.
"I don't like <person or organization>'s doing". If that's the whole message, all you probably did so far is complaining or crying. Sweet tears.
Stuff like that can happen to any person, just like any person makes mistakes.
Mistakes are made to learn from them. If you realize realize and accept your mistakes others may do so as well and forgive you.
But we are he towards this idiotic trend where people just can swallow their pride even for microscopic things. They instead push their pride to higher levels of ignorance, blaming other people, l(ying)mfao, creating black holes of density in the process. Makes me wonder whether their real motive is an inside bet on who can get the most people to kill them selves by face palming.
Most of my life I have been fairly protected against these people, besides some spikes of incompetence, but recently the have invaded 2 areas in my world that make the world somewhat less of a pain. Programming and the internet culture.
Yes, I'm talking about that master / slave BS renaming and article 11 and 13.
The remaking itself isn't really the problem, but rather the context. This was basically a show of power for the self proclaimed "social justice warriors" or SJW for short.
The fact that this madness has spread. That's what worries me. To me it feels like the first zombie has spawned.
Then we have this corrupted piece of incompetent shit, called Axel Voss, and other old farts.
They live in a galaxy far away from reality, somewhere in the European Parlament, making laws they don't know shit about, regulating things they know shit about.
All in the name of the people of the EU of course. And by people we obviously talk about the money.
I can honestly not think of another reason, after reading the replies Voss and his party gave on Twitter regarding the shit they pulled off.
Well, at least none that doesn't involve some firm of brain death.
For now I'll show them as much as possible how much I despise / reject them. Currently playing with the thought of some kind (social media?) website were posts from other sites or actions in general can be rated only with "Fuck you"s.
Given these articles, I should not have them hosted in an European country though 😅.
Almost hitting that 5k character limit 😰1 -
Assumption. Screens are wider and wider, bigger and bigger.
Solution 1: Make more one-page layout sites like there is no space on sides.
Solution 2 support solution 1: Make everything 2-3 times bigger so there can't be space on sides.
Solution 3 support solution 2: If there is some free space, call it neccessary white space.
It's a trend i see around me. Sometimes i don't get it. More than 1 year i have no side menu in designes what i coded. -
A system to build note-taking systems. tatatap dot com.
It’s the most successful for a few reasons: it got launched, people find it useful, but most importantly it’s been fun and continues to be fun to work on.
I think the fun-to-make factor is massively underestimated as a success indicator. Working on the right product (whatever that means) that is unenjoyable is like using an amazing computer with a broken keyboard. It’s never going to work.
Sure, with any project there’s annoying stuff, but it’s the trend overall. Is the core functionality fun to work on?
In the case of Tap the core component is a notation parser, open sourced called sowhat, github dot com/tatatap-com/sowhat
That was super fun to make and learn about lexing and parsing. It’s pretty far along but there’s still a lot I’m planning to add. -
Why do modern Europeans like to wear wigs?
The prevalence of wigs is closely related to the social life conditions at that time. Because in the 17th century, Europe, it was very inconvenient for people to bathe and wash their hair. Louis XIV, the famous Sun King, took only seven baths in his life. Not taking a long bath and shampoo, it is easy to breed parasites, especially hair, hair thick, often sweat, it is easy to grow lice. The best way to solve this problem is to cut the hair short or shaved, but the hair is cut short or shaved, and can not reflect the identity of aristocrats, it is better to wear a wig, have the best of both worlds.
In addition to the aristocracy as a fashion, the real problem for a wig to become a status symbol, is that the wig is expensive and the average person cannot wear it. In the 17th century, the wig was very elaborate. At that time, there was no machine production, so it depended on labor. A skilled craftsman needed a few days to make a wig. A judge's wig costs £1,800, and a regular wig costs £300. This money is a huge expense today, not to mention Western Europe before the Industrial Revolution. Therefore, wearing wigs is not something that ordinary people can afford. And at that time, the wig was quite bulky, also uncomfortable to wear, often working people naturally will not wear.
In addition to being expensive and inconvenient to wear, the embellishment and maintenance of wigs are also quite cumbersome. The 18th-century wig often had some pollen and some paint added. Pink wigs are easy to drop powder, and they are difficult to take care of. So, it is naturally not favored by ordinary people. By the late 18th century, young men simply added powder to their hair. The wigs worn by women were large and striking, but they were heavy and contained wax, powder and other ornaments, becoming a sign of luxury.
However, with the beginning of the Industrial Revolution in the middle of the 18th century. Natural hair without wigs is slowly being accepted by more people. In Goethe's masterpiece, "The Trouble of the Young Witt," Witt's natural hair triggered a natural fashion trend at the time. After the outbreak of the French Revolution, the revolutionaries tried to establish an equal society, eliminate class differences, and the wigs representing their status were naturally among the objects of changing customs.
In addition, in 1795, the British government began to tax the hair fans, which hit the wig and hair fan fashion, and began to decline in the 19th century. By the 19th century, the wigs became smaller and grave. In France, wigs are no longer a status symbol. But wigs remained as a status symbol for some time. After the French Revolution, French wigs, which no longer became a symbol of status, were associated with professional prestige. Some industries and fields use wigs as part of their professional clothing, such as judges and lawyers. This habit continues to this day. Judges and lawyers in the Commonwealth wear wigs in court or at major ceremonies, a tradition in previous British colonies, but it makes them a mark of colonial rule.
The popularity of a generation of fashion, it must have its historical background, once1