Join devRant
Do all the things like
++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatar
Sign Up
Pipeless API
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple API
Learn More
Search - "constructed"
-
Last Friday, because of a stupid ass shitty engineer who constructed a footpath in the most shitty way I Sprained my ankle in one of the dents in it. But the story about this pizza picture is that my friend and I had planned a trek on a Saturday which obviously got cancelled. He saw how pissed I was he went online on Domino's website, changed input in a JSON file on one of Domino's wheel spinner which got me the maximum discount available there and I bought the pizza. He did that multiple times. Some Relief there. 🥂🍕
#Pizza_makes_everything_better11 -
In Russia we have a huge techno-nazi community. They often can be found on some programming forums and a website called Habr.
They’ll shame you if you’re a web developer and don’t write in Asm or C. They spread toxic memes and insulting “stupid humanitarians”. One particular guy constructed the whole ideology that is focused on technological enthusiasts being the “master race” and implies recycling non-technical people in so-called “bioreactors” for energy.
Please don’t be like them.35 -
I was young and stupid. Remember floppy disks? Yeah, we still had them when I was studying. Went to a computer cafe, rented a PC and DOS was already booted up (I'm an old fucker). I didn't want to reboot because PC rental was metered. I inserted my floppy disk and got infected by a virus that deleted my work. No git back then and my backup was on the same disk (fuck me). Back home, rewrote the whole thing from memory.
I got mad and wanted revenge. De-constructed a floppy disk, replaced the magnetic media with sandpaper and went to each and every PC on that computer cafe.
It was closed for day.3 -
650 line hand constructed xml file missing a tag. Took me 3 hours to find. I looked at the section at least 3 times but my brain would not pick out the mistake. Online cal validators all pointing me to the wrong places.
It takes persistence to be a developer so pat yourselves on the back.5 -
Just had an old coworker from a previous job send me some stuff for a php script he was having issues with.
There was too much glory in what he was trying to do: mixing php inside of jquery code, not using strict types would have prevented like 10 issues he was having on his script on another portion, mixing headers, weirdly named variables, poorly constructed, reused db connections, 0 oop or proper dependency management in his code, horrible use of sessions and cookies, O (n²) logic all over the place.
But the cake.....are y'all ready for it? It was code screenshots, not even of just the section, no, the full page, from a windows machine (to make it better he is hosting the application on an IIS server and his configuration was not properly set) but I digress, back to the cake:
He was writing his code inside of wordpad :P
FUCKING WORDPAD
I just politely told him that I was busy at the moment and happily ignored him. Dude is not a good person to begin with imo, for example, he brought the subject of homosexuality during one of our talks after he saw me talking to my bf, who just so happens to be gay, his statement was "I do not understand how there can be gay people when there are women that are so hot"
My comeback was "I do not understand how we can be heterosexual when there are some really attractive dudes out there, see how stupid your logic sounds? attractiveness is not the basis for homosexuality ye dipstick" he let it go after that, but close minded people like that are not really my cup of tea.14 -
Halloween is coming so i made this constructed unicorn mask with a paper mache base with elwire. Does it require any coding skills nope, but i bet people are goin to be suprised that i know how to build wireframes and papermache !7
-
A bug which takes 6 days to be found/fixed. Screw you financial data. 👹
🏀on the good side: whole class was de-constructed and re-factored + no other bugs were found! 😀🔥1 -
My friend: I built a 3D printer and coded it to self calibrate at startup and connect to my encrypted remote server through VPN to securely retrieve Cad files I constructed and start itself printing when everything is ready or alert if theres an issue
Me: aren't you less than a year older than me?
This was when we were underclassmen in college a few years ago 😂 turns out the dudes just a savant5 -
Sent a fully constructed sql statement to someone expecting at least 4 rows however received reply "the result is empty" with a screenshot of empty result set from sql-developer. I kept cross-checking the where clause thinking I mixed something up.
After a few back and forth emails suddenly noticed the screenshot I received initially and I see all of the strings in where clause are lowercase. I reference my version and it is correct. When I asked her why are the strings in sql lowercase and that if she has tried the exact sql I sent in the email, the response "I didn't think it mattered what case the sql was in".
I am lost for words. The worse part is, this is someone who is supposed to go on site as part of their job and help clients setup, explain and train how the software works. This includes explaining how software intreacts with database tables 🤐8 -
QA personal voice assistant that runs locally without cloud, it’s like never ending project. I look at it from time to time and time pass by. Chat bots arrived, some decent voice algorithms appeared. There is less and less stuff to code since people progress in that area a lot.
I want to save notes using voice, search trough them, hear them, find some stuff in public data sources like wikipedia and also hear that stuff without using hands, read news articles and stuff like that.
I want to spend, more time for math and core algorithms related to machine learning and deep learning.
Problem is once I remember how basic network layers, error correction algorithms work or how particular deep learning algorithm is constructed and why is that, it’s already a week passed and I don’t remember where I started.
I did it couple of times already and every time I remember more then before but understanding core requires me sitting down with pen and paper and math problems and I don’t have time for that.
Now when I’m thinking about it - maybe I should write it somewhere in organized way. Get back to blogging and write articles about what I learned. This would require two times the time but maybe it would help to not forget.
I’m mostly interested in nlp, tts, stt. Wavenet, tacotron, bert, roberta, sentiment analysis, graphs and qa stuff. And now crystallography cause crystals are just organized graphs in 3d.
Well maybe if I’m lucky I retire in the next decade or at least take a year or two years off to have plenty of time to finish this project. -
There is a constructed language called Tokipona. It was made to be the easiest language, and it only contains like 137 words or so.
This language is a perfect demonstration of “expressivity”. Tokipona is not very expressive. Before you know it, your sentence is obscenely long, and you didn't even convey the full meaning of what you wanted to say.
It's also the case with “easy” programming languages and frameworks. Code quantities rise exponentially, and the more code you have, the more bugs there are. There is no magic. And then you have to debug it. Not so easy, huh?10 -
Ithkuil is beautiful.
A constructed language to eliminate double meanings. A lot of choice on concepts, the info packed as densely as possible.15 -
Am I a hack? Like yeah I complain about technology left right and center, this sucks, that sucks, what fucking moron wrote this?! These days I do write my own alternatives (which usually work surprisingly well). But for what? And was I really in a position to complain about those other things? Impostor syndrome, it's so annoying...
Oh and also, is it really all worth it? I like retro tech and so I do have a fair interest in the history of technology. Say between VHS and Beta, sure VHS was superior in practice and won the video cassette war, but Beta machines were seemingly better constructed. VHS won because it did just enough. Perhaps the same is true for software? Overengineering, is it poor engineering?
Anyone can build a bridge if the budget is unlimited and it can take a lifetime to construct. But part of engineering is making a bridge that'll just barely stand and be finished in a few years. I've been working on my own Linux distro since August last year and am not even close to finishing it. Chances are that it'll take several years. Perhaps I've been looking at the problem the wrong way all along?1 -
Last week me and my friend have been changed from a legacy PHP project to new Ruby on Rails-based setup. What, in first instance, looked like a great improvement, now becomes a nightmare.
All this convention-over-configuration is awesome - but only if you already know the conventions, or if somebody told'em to you.
And everything is going even more out of control because the damn project is based upon Spree gem and several other extensions, that MUST be changed to meet out company needs.
I'm getting really mad with all this pressure. Ruby seems to be a great language, but I'd rather be working with Laravel. Its overall organization, the centralization of CLI commands in artisan, and the astoundingly clear, eloquent, direct and well-designed documentation made my adoption curve there a little more pleasant.
I mean, legacy PHP systems are awful, but Laravel framework sounds way more easy-to-learn and well-constructed when compared to rails.
But given all this nightmare, I really want to be proved the opposite.1 -
More from my big black book of ai and neuroscience:
I think if trace theory is true to any degree it would go some distance in explaining phenomenal consciousness, assuming I haven't misunderstood anything.
In fuzzy trace theory (FTT) it is posited that people form two types of mental representations about a past event:
*verbatim traces: detailed representations of a past event.
*gist traces: fuzzy representations of a past event.
People can reason with verbatim *and* gist traces but prefer gists.
*vision was suggested to work similarly in 1999. With human vision, two processes could be used: one that aggregates local receptive fields and one that parses the local receptive spatial field. It was suggested that people used prior experience, gists, to decide which dominates a perceptual decision.
Gist processes form representations of events, semantic details, where verbatim reinstates the context found in the surface details of an event.
__notes__
Parallel storage: asserts encoding/storage of verbatim/gist traces operate in *parallel*, not in serial.
I like to think of verbatim traces as databases, and gists as queries constructed by recognition.
Several studies have found that the meaning (gist) of an item is encoded even *before* the surface details (verbatim).
This might be important as a survival mechanism but should not be taken to mean strictly that gists are formed wholly *without* details or important and recognizable features of the item in question. It may well be for high level el processing and classification efficiency this may be an important reprocessing step, in the same way that many functions of the brain are duplicated throughout.5 -
Every day and especially at night in my bed, I go through different websites that talk about the web design and development. It's very important to me .. To learn and always be informed of trends.
What is the website that you like the most? The one you read every day and you think to be the only one to know ? Mine is : Website Deconstruction ( http://websitedeconstructions.com ) . A website that dissects others to understand how they are constructed and how they work.
A really good one.3 -
I'm still on a regular basis reminded of how I might be wrong despite the absolute certainty in how obviously wrong the other person is.
Lately I've been working on setting up this API with a fairly intricate database integration. One request can lead to multiple db calls if we're not careful, so we have been polishing up the implementation to guard against ddosing ourselves and dealing with thread-unsafe concurrency.
Someone on the team could happily report that they got rid of all async use so there should no longer be threading issues. "You mean it all runs sync now?" "I guess. It works at least".
I'm just internally pulling a surrender cobra. If this was pre-dev me I would have let him and everyone know what a stupidpants he is and that I thought he had some experience in api development. But let's not make an exception to the rule; I might be wrong. I mean I'm not, but let's pretend I could be. Let's pull down the changes and maybe set up a minimal example to demonstrate how this is a bad idea.
Funny story. He got rid of explicit calls to the database entirely. When resolving data, the query is instead constructed virtually and execution is deferred until the last step. Our functions are sync now because they don't call the database, and threading isn't an issue since there's only one call per request context.
Thank god I've learned to keep my mouth shut until I can prove with absolute conclusive certainty that they are wrong. Here's to another day of not making an ass of myself. -
Hey girl are you into PDA? Because I've constructed a Pushdown Automata that can generate binary palindromes.1
-
University labs. I’ve been lucky enough to have 2 brand new computing labs constructed during my time here. Ultra wide QHD USB-C monitors 😍
-
So burnt out. Zomg. The typos... the poorly constructed tests... the clients daily changes and the urgency of it all... my test created an unecessary object that had me looking at all the wrong things. My own dumbness. From overwhelmed import haste
-
I wrote a small crate that does unsafe operations, please help me verify its soundness: https://github.com/lbfalvy/bound
(Also I think you'll like it, I'm solving a fairly abstract problem that's not possible in safe Rust)
It's essentially a struct that ties together a heap reference and a struct that's constructed from it. The main use case is to return lock guards derived from Arc<Mutex> but it's defined in a very abstract way intentionally because I'm using Marc from mappable-rc and async-std's RwLock and I didn't want this to depend on either crate.
It actually has no dependencies apart from STD (I think this one may be unavoidable) -
None::<()>
I hope as never (`!`) gets stabilized in Rust, they come up with a clever heuristic to set it as the default value of generic parameters that cannot be deduced because the variants using them are never constructed. -
Once upon a time I offhandedly suggested a few beautiful things that were constructed
People who were taking responsibility for the ideas who do not understand delight simple Innocent delight and a desire to see what I created before had some miniatures cheaply constructed
The desire is for things that endure in memory
All they wanted was to mechanically and immediately work on something else like it was some kind of souless side show
All I wanted was a few measly days of rest and to look over the small shadows of my ideas
This is why 99% of all new ideas suck2 -
so how about to get past that big monster at the end of time, which has been thus constructed into a loop by a very large interlinking collection of lying idiots.4
-
The Turing Test, a concept introduced by Alan Turing in 1950, has been a foundation concept for evaluating a machine's ability to exhibit human-like intelligence. But as we edge closer to the singularity—the point where artificial intelligence surpasses human intelligence—a new, perhaps unsettling question comes to the fore: Are we humans ready for the Turing Test's inverse? Unlike Turing's original proposition where machines strive to become indistinguishable from humans, the Inverse Turing Test ponders whether the complex, multi-dimensional realities generated by AI can be rendered palatable or even comprehensible to human cognition. This discourse goes beyond mere philosophical debate; it directly impacts the future trajectory of human-machine symbiosis.
Artificial intelligence has been advancing at an exponential pace, far outstripping Moore's Law. From Generative Adversarial Networks (GANs) that create life-like images to quantum computing that solve problems unfathomable to classical computers, the AI universe is a sprawling expanse of complexity. What's more compelling is that these machine-constructed worlds aren't confined to academic circles. They permeate every facet of our lives—be it medicine, finance, or even social dynamics. And so, an existential conundrum arises: Will there come a point where these AI-created outputs become so labyrinthine that they are beyond the cognitive reach of the average human?
The Human-AI Cognitive Disconnection
As we look closer into the interplay between humans and AI-created realities, the phenomenon of cognitive disconnection becomes increasingly salient, perhaps even a bit uncomfortable. This disconnection is not confined to esoteric, high-level computational processes; it's pervasive in our everyday life. Take, for instance, the experience of driving a car. Most people can operate a vehicle without understanding the intricacies of its internal combustion engine, transmission mechanics, or even its embedded software. Similarly, when boarding an airplane, passengers trust that they'll arrive at their destination safely, yet most have little to no understanding of aerodynamics, jet propulsion, or air traffic control systems. In both scenarios, individuals navigate a reality facilitated by complex systems they don't fully understand. Simply put, we just enjoy the ride.
However, this is emblematic of a larger issue—the uncritical trust we place in machines and algorithms, often without understanding the implications or mechanics. Imagine if, in the future, these systems become exponentially more complex, driven by AI algorithms that even experts struggle to comprehend. Where does that leave the average individual? In such a future, not only are we passengers in cars or planes, but we also become passengers in a reality steered by artificial intelligence—a reality we may neither fully grasp nor control. This raises serious questions about agency, autonomy, and oversight, especially as AI technologies continue to weave themselves into the fabric of our existence.
The Illusion of Reality
To adequately explore the intricate issue of human-AI cognitive disconnection, let's journey through the corridors of metaphysics and epistemology, where the concept of reality itself is under scrutiny. Humans have always been limited by their biological faculties—our senses can only perceive a sliver of the electromagnetic spectrum, our ears can hear only a fraction of the vibrations in the air, and our cognitive powers are constrained by the limitations of our neural architecture. In this context, what we term "reality" is in essence a constructed narrative, meticulously assembled by our senses and brain as a way to make sense of the world around us. Philosophers have argued that our perception of reality is akin to a "user interface," evolved to guide us through the complexities of the world, rather than to reveal its ultimate nature. But now, we find ourselves in a new (contrived) techno-reality.
Artificial intelligence brings forth the potential for a new layer of reality, one that is stitched together not by biological neurons but by algorithms and silicon chips. As AI starts to create complex simulations, predictive models, or even whole virtual worlds, one has to ask: Are these AI-constructed realities an extension of the "grand illusion" that we're already living in? Or do they represent a departure, an entirely new plane of existence that demands its own set of sensory and cognitive tools for comprehension? The metaphorical veil between humans and the universe has historically been made of biological fabric, so to speak.7 -
You know in time all that will be left of them is maybe the idea that they were all whores and maybe people will feel sorry for them long after their dead
And that would be good because it would encourage a sense of humanity in future generations which being the exact opposite of what they want would be part of the sweet revenge
I think that splitting them off from the group that does all this creepy shit would also be a nice alteration to history
It would allow the young to despise the one while not falling victim to the propaganda they use to try to humanize themselves to cause other people grief and trauma down the road and would not allow them to falsely portray all people stuck in their line of work as the same kind of garbage trash that has no other use
Wouldn't that be lovely ? All the mind numbing buildup of chomo trash you people constructed torn down meaning lost and the ambiguous nature of much of it portrayed as it was portrayed as ordinary sex games and the like and adventures being left behind to delight people you'd all victimize in future generations ? All your wasted fucked up lives reduced to zero. Just like you all forced on so many others ?
Reverse pronouns if this isn't making sense since everyone knows you people speak English but just act like retards.
In time the world will heal
End of story
The perfect formula for screwing over younger straighter more innocent and good natured if lusty and angry people will no longer work and your fucked up abuses will disappear and noone will remember any of your names just like you creatures tried by stealing everything decent people created and passing it off as your own. And your dumb code will be as nonsensical then as it is now
Glorious
At least in the long run there is that as this evil is purely self destructive8