Do all the things like ++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatarSign Up
From the creators of devRant, Pipeless lets you power real-time personalized recommendations and activity feeds using a simple APILearn More
Search - "containers"
I’m kind of pissy, so let’s get into this.
My apologies though: it’s kind of scattered.
For @Root? Fucking never.
Maybe if I wanted to be a business major my mother might have cared. Maybe the other one (whom I call Dick because fuck him, and because it’s accurate) would have cared if I suddenly wanted to become a mechanic. But in both cases, I really doubt it. I’d probably just have been berated for not being perfect, or better at their respective fields than they were at 3x my age.
Support being a dev?
Not even a little.
I had hand-me-down computers that were outmoded when they originally bought them: cutting-edge discount resale tech like Win95, 33/66mhz, 404mb hd. It wouldn’t even play an MP3 without stuttering.
(The only time I had a decent one is when I built one for myself while in high school. They couldn’t believe I spent so much money on what they saw as a silly toy.)
Using a computer for anything other than email or “real world” work was bad in their eyes. Whenever I was on the computer, they accused me of playing games, and constantly yelled at me for wasting my time, for rotting in my room, etc. We moved so often I never had any friends, and they were simply awful to be around, so what was my alternative? I also got into trouble for reading too much (seriously), and with computers I could at least make things.
If they got mad at me for any (real or imagined) reason (which happened almost every other day) they would steal my things, throw them out, or get mad and destroy them. Desk, books, decorations, posters, jewelry, perfume, containers, my chair, etc. Sometimes they would just steal my power cables or network cables. If they left the house, they would sometimes unplug the internet altogether, and claim they didn’t know why it was down. (Stealing/unplugging cables continued until I was 16.) If they found my game CDs, those would disappear, too. They would go through my room, my backpack and its notes/binders/folders/assignments, my closet, my drawers, my journals (of course my journals), and my computer, too. And if they found anything at all they didn’t like, they would confront me about it, and often would bring it up for months telling me how wrong/bad I was. Related: I got all A’s and a B one year in high school, and didn’t hear the end of it for the entire summer vacation.
It got to the point that I invented my own language with its own vocabulary, grammar, and alphabet just so I could have just a little bit of privacy. (I’m still fluent in it.) I would only store everything important from my computer on my only Zip disk so that I could take it to school with me every day and keep it out of their hands. I was terrified of losing all of my work, and carrying a Zip disk around in my backpack (with no backups) was safer than leaving it at home.
I continued to experiment and learn whatever I could about computers and programming, and also started taking CS classes when I reached high school. Amusingly, I didn’t even like computers despite all of this — they were simply an escape.
Around the same time (freshman in high school) I was a decent enough dev to actually write useful software, and made a little bit of money doing that. I also made some for my parents, both for personal use and for their businesses. They never trusted it, and continually trashtalked it. They would only begrudgingly use the business software because the alternatives were many thousands of dollars. And, despite never ever having a problem with any of it, they insisted I accompany them every time, and these were often at 3am. Instead of being thankful, they would be sarcastically amazed when nothing went wrong for the nth time. Two of the larger projects I made for them were: an inventory management system that interfaced with hand scanners (VB), and another inventory management system for government facility audits (Access). Several websites, too. I actually got paid for the Access application thanks to a contract!
To put this into perspective, I was selected to work on a government software project about a year later, while still in high school. That didn’t impress them, either.
They continued to see computers as a useless waste of time, and kept telling me that I would be unemployable, and end up alone.
When they learned I was dating someone long-distance, and that it was a she, they simply took my computer and didn’t let me use it again for six months. Really freaking hard to do senior projects without a computer. They begrudgingly allowed me to use theirs for schoolwork, but it had a fraction of the specs — and some projects required Flash, which the computer could barely run.
Between the constant insults, yelling, abuse (not mentioned here), total lack of privacy, and the theft, destruction, etc. I still managed to teach myself about computers and programming.
In short, I am a dev despite my parents’ best efforts to the contrary.31
So i heard that docker is now a thing and decided to put my servers into containers as well (see image)...but somehow I still don't understand how is this supposed to simplify deployment, vertical scaling and so on...? :D6
enabling firewall on a vps to secure my docker containers and forgetting to add openssh to allowed list --> ssh blocked 😃🔫24
I'm not even that old and I've had it with young cocksure, full of them self language/environment evangelists.
- "C# is always better than Java, don't bother learning it"
- "Lol python is all you need"
- "Omg windows/linux/mac sucks use this instead"
The list goes on really, at some point you have got to realize that while specialization is great, you have to learn a little bit of everything. It broadens you horizon a lot.
Yea, C# does some nifty stuff, but Java does too, learn both. Yea I'm sure Linux is better for hosting docker containers, but your clients are on mac or windows, learn to at least navigate and operate all three etc. Embrace knowledge from all the different tech camps it can only do you good and you will be so much more flexible and employable than your close minded peers :)
Hell even PHP has a lot to teach us (Even more than just to be a bad example, har har)9
That awkward moment when I was able to run three docker containers on a 512MB server:
1. DotNet core web service
BUT I cannot run:
1. NodeJs web service
2. MongoDB container
Spent two hours configuring the damn server to get hit by this T_T14
Soooo my little encryption tool makes progress. <3
After a short break from development, we had our first successful loaded container yesterday!
- Protocoll is working
- We can create containers and store/copress files in it
- we're awesome
- I love it
- you are awesome, too!
(Loaded containers will be inaccessible for movement to different directories while our tool is open)49
So this client wanted a demo on Dockers. So I gave the demo with some microservices running on different containers. Later the clients come back and say, "Docker is good. But please fit all the microservices in one container." I say but that defeats the purpose of microservices. But no, the client say. I tried explaining but no is no. Shit!! Fine! Have it your way!!5
This rant is particularly directed at web designers, front-end developers. If you match that, please do take a few minutes to read it, and read it once again.
Web 2.0. It's something that I hate. Particularly because the directive amongst webdesigners seems to be "client has plenty of resources anyway, and if they don't, they'll buy more anyway". I'd like to debunk that with an analogy that I've been thinking about for a while.
I've got one server in my home, with 8GB of RAM, 4 cores and ~4TB of storage. On it I'm running Proxmox, which is currently using about 4GB of RAM for about a dozen VM's and LXC containers. The VM's take the most RAM by far, while the LXC's are just glorified chroots (which nonetheless I find very intriguing due to their ability to run unprivileged). Average LXC takes just 60MB RAM, the amount for an init, the shell and the service(s) running in this LXC. Just like a chroot, but better.
On that host I expect to be able to run about 20-30 guests at this rate. On 4 cores and 8GB RAM. More extensive migration to LXC will improve this number over time. However, I'd like to go further. Once I've been able to build a Linux which was just a kernel and busybox, backed by the musl C library. The thing consumed only 13MB of RAM, which was a VM with its whole 13MB of RAM consumption being dedicated entirely to the kernel. I could probably optimize it further with modularization, but at the time I didn't due to its experimental nature. On a chroot, the kernel of the host is used, meaning that said setup in a chroot would border near the kB's of RAM consumption. The busybox shell would be its most important RAM consumer, which is negligible.
I don't want to settle with 20-30 VM's. I want to settle with hundreds or even thousands of LXC's on 8GB of RAM, as I've seen first-hand with my own builds that it's possible. That's something that's very important in webdesign. Browsers aren't all that different. More often than not, your website will share its resources with about 50-100 other tabs, because users forget to close their old tabs, are power users, looking things up on Stack Overflow, or whatever. Therefore that 8GB of RAM now reduces itself to about 80MB only. And then you've got modern web browsers which allocate their own process for each tab (at a certain amount, it seems to be limited at about 20-30 processes, but still).. and all of its memory required to render yours is duplicated into your designated 80MB. Let's say that 10MB is available for the website at most. This is a very liberal amount for a webserver to deal with per request, so let's stick with that, although in reality it'd probably be less.
10MB, the available RAM for the website you're trying to show. Of course, the total RAM of the user is comparatively huge, but your own chunk is much smaller than that. Optimization is key. Does your website really need that amount? In third-world countries where the internet bandwidth is still in the order of kB/s, 10MB is *very* liberal. Back in 2014 when I got into technology and webdesign, there was this rule of thumb that 7 seconds is usually when visitors click away. That'd translate into.. let's say, 10kB/s for third-world countries? 7 seconds makes that 70kB of available network bandwidth.
Web 2.0, taking 30+ seconds to load a web page, even on a broadband connection? Totally ridiculous. Make your website as fast as it can be, after all you're playing along with 50-100 other tabs. The faster, the better. The more lightweight, the better. If at all possible, please pursue this goal and make the Web a better place. Efficiency matters.9
DISCLAIMER: UNPOPULAR OPINION
I'm tired of the Linux community, they effectively discourage me of taking part in any discussion online
I'm currently making Windows-only soft, some game stuff, some legacy DirectX stuff you got it.
Everytime I go online, this shitty pattern happens, when I stumble upon a problem in project I don't know how to fix and I ask for help
These are responses
- HA, HA, WINDOWS BAD, HA, HA, GET REAL SYSTEM
- In Linux, we can do X too. I mean it has 4x less functionality and way shittier UX and is even harder to implement but it can probably work on too Linux, so it's better, yes, just move to Linux
- btw you didn't like Linux before? Try this distro man, it's better <links random distro>
Is there anything valuable in the Linux community? I feel like these people don't like Linux anyway, they just hate Windows. Every opinion, tip is always opinion based. Anyone who works on internals knows how much better and how well thought is Windows kernel compared to Linux kernel. Also, if someone unironically uses Linux distro on desktop PC then he's a masochist because desktop Linux is dieing. So many distros ceased work only this year.
Is it a good tool for servers and docker containers? I don't have my head stuck up my ass to admit that yes, it's much better than Windows here.
This community got me stressed right now, I fear that when I go to bathroom or open my microwave there's gonna be a Linux distro recommendation there
I'm just amazed what 512MB of RAM can do :O
That's htop from my VPS I feel sorry for the CPU though.
It is running three docker containers:
1. Dotnet Core
Are you using socat?
Any interesting use case you would like to share?
I am using it to create fake / proxy docker containers for network testing.7
Now seriously, WHAT THE FUCK???
Every single time I have to work with people from a particular country [you have one guess. Yepp, that's the one], I see A-FUCKING-LOOOOOOT of manual work?!?
"can you reboot the server?"
-"sure, let me help you, sir" <20 minutes later> "done"
"can you unlock my account?"
-"yes, just a moment sir" <20 minutes later> "please check now"
"can you restart this environment w/ 200 instances?"
-"yes sir, let me check" <6 hours later> "please check now"
"you've missed 18 containers"
-"oh okay sir, will restart them now" <2hours later> "please check now"
[I am already OoO]
why is it that every time I have to work with you guys I am the one who is automating shit. How come you never think of/do any automata? You are fucking technitians, you should know how. WHY DO YOU ENJOY CLICKING ALL-DAY-LONG????
I'm serious. Why??? I'm struggling to understand...25
Junior dev at my workplace keeps telling me how efficient docker is.
He decided to solve his latest task with a containers in swarm mode.
As expected, things went sideways, and I had the joy of cleaning up behind him.
A couple of days later, I noticed that I was running low on disk space - odd.
Turns out docker was eating up some 60 GB with a bunch dangling images - efficient is a funny term for this.17
Manager: "Can we get an accurate report on how many containers we have on the Kubernetes cluster?"
Me: "Well not really since Kubernetes is designed to be dynamic and agile with the number of resources and containers being created and deleted being subject to change at a moment's notice."
Manager: "I want numbers"
Me: "Okay well if we look at a simple moving average over time, we can see how the number of containers changes and then grab a rough answer from that"
Manager: "These numbers look a little round, are you sure these are exact?"
I'm going to throw myself into a pile of used heroin needles and hope i get stuck with whatever the hell this guy has to somehow be a manager while also being this retarded.15
Go home grafana, you're drunk....
Maybe I shouldn't run 50 containers on a system with 2 cores and 6 GB RAM.4
Our client decided to save some $$. At the end of each business day teams downscale their environments before leaving and the next day scale them up in the morning to start working.
The idea is not bad, but they are a bit too ignorant to the fact that some environments are exceeding AWS APIs limits already (huge, HUGE accounts, huge environments, each env easily exceeding /26 netmask, not even taking containers into account). Sooo... scaling up might take a while. Take today for example:
- come in to the office at 7
- start scaling up
- have lunch
- ~15:00 scaleup has finished
- one component is not working, escalating respective folks to fix them
- ~17:00 env is ready for work
- 17:01 initiate scaledown process and go home
Sounds like a hell of a productive day!!!
FUCK capitalist greed! I have befallen to their tricks once again. The daily dosage on my gummy vitamins was three a day but the total gummies in the container wasn't divisible by three so I had to buy three containers and eat one from each per day!21
Okay guys, this is it!
Today was my final day at my current employer. I am on vacation next week, and will return to my previous employer on January the 2nd.
So I am going back to full time C/C++ coding on Linux. My machines will, once again, all have Gentoo Linux on them, while the servers run Debian. (Or Devuan if I can help it.)
So what have I learned in my 15 months stint as a C++ Qt5 developer on Windows 10 using Visual Studio 2017?
1. VS2017 is the best ever.
Although I am a Linux guy, I have owned all Visual C++/Studio versions since Visual C++ 6 (1999) - if only to use for cross-platform projects in a Windows VM.
2. I love Qt5, even on Windows!
And QtDesigner is a far better tool than I thought. On Linux I rarely had to design GUIs, so I was happily surprised.
3. GUI apps are always inferior to CLI.
Whenever a collegue of mine and me had worked on the same parts in the same libraries, and hit the inevitable merge conflict resolving session, we played a game: Who would push first? Him, with TortoiseGit and BeyondCompare? Or me, with MinTTY and kdiff3?
Surprise! I always won! 😁
4. Only shortly into Application Development for Windows with Visual Studio, I started to miss the fun it is to code on Linux for Linux.
No matter how much I like VS2017, I really miss Code::Blocks!
5. Big software suites (2,792 files) are interesting, but I prefer libraries and frameworks to work on.
For future reference, I'll answer a possible question I may have in the future about Windows 10: What did I use to mod/pimp it?
1. 7+ Taskbar Tweaker
3. Classic Start (Now: Open-Shell-Menu)
Enhanced text editor I like a lot more than notepad++. Aaaand it has a "vim-mode". 👍
Three way diff viewer, that can resolve most merge conflicts on its own. Its keyboard shortcuts (ctrl-1|2|3 ; ctrl-PgDn) let you fly through your files.
8. Link Shell Extensions
Support hard links, symbolic links, junctions and much more right from the explorer via right-click-menu.
Neither as beautiful as Conky, nor as easy to configure or flexible. But it does its job.
Of course this wasn't everything. I also pimped Visual Studio quite heavily. Sam question from my future self: What did I do?
1 AStyle Extension
2 Better Comments
Simple patche to make different comment styles look different. Like obsolete ones being showed striked through, or important ones in bold red and such stuff.
4 Atomineer Pro Documentation
Alright, it is commercial. But there is not another tool that can keep doxygen style comments updated. Without this, you have to do it by hand.
5 Highlight all occurrences of selected word++
Select a word, and all similar get highlighted. VS could do this on its own, but is restricted to keywords.
6 Hot Commands for Visual Studio
This ingenious invention colorizes brackets (aka "Rainbow brackets") and makes their inner space visible on demand. Very useful if you have to deal with complex flows.
Come on! 2018 and Visual Studio still outputs monochromatically?
That's it, folks.
No matter how much fun it will be to do full time Linux C/C++ coding, and reverse engineering of WORM file systems and proprietary containers and databases, the thing I am most looking forward to is quite mundane: I can do what the fuck I want!
Being stuck in a project? No problem, any of my own projects is just a 'git clone' away. (Or fetch/pull more likely... 😜)
Here I am leaving a place where gitlab.com, github.com and sourceforge.net are blocked.
But I will also miss my collegues here. I know it.
Well, part of the game I guess?7
Sometimes I daydream about just selling out and doing the whole diversity and women who code kind of thing. Apply for the big tech companies and work my way towards becoming one of the faces of their bullshit company. I'm gonna get a long wig, tan myself brown, and wear subtle native shit like a carabao brooch or some Cheongsam-inspired work suit. And when there's a big work event or party, I'm gonna go full on White Rose from Mr Robot, smiling and shaking people's hands like I give a shit. Gonna try to look wholesome and innocent, all I really wanted was a chance to get into tech as a humble being.
I'm gonna practice a strong accent and I would tell humble stories about where I came from, that god sent me here, and about all the struggles I had getting into the field as a woman. I'm gonna step up to the podium and say shit like, "Women can do it too, we can kuh-hode!" Then I'm gonna stir up some shit like, "Why is the terminal black?" Gonna make my voice and hands shake from intense emotion (visible anger, almost crying), "Why is something that we use to do our bidding black? Are we not over the years of slavery in the past? What does a terminal represent? Slavery. You know what else is terminal? Cancer. Slavery and racism are the cancers of our society. We must stop this. Terminals should be in all colors and it shouldn't be black by default."
1. Point out a non-malicious thing and add malice to it. The more racist and sexist, the better.
2. Mention something bad that is not related to it at all.
3. Make them seem related in an attempt to make a point.
4. Fail to make a point.
5. Say "we must stop this".
I'm gonna look straight into people's eyes with a serious face while cry-laugh-cringing internally. "As a brown woman, this makes me sad. I have Kubernetes skills, you know and it hurts me." If I catch some people laughing even though I'm also about to fart from internal laughter, I will point at the whitest male of them all. "You, do you think this is funny? Because this is my life, as well as millions of other people like ME!"
Then I'm gonna curse all of the white men who obviously only got into this field through violence. Yes, Mike, I know you pointed a gun at the HR to get in. What else could it be? All you know is violence because testosterone! You don't really know Kubernetes, I do, I set it up all the time for my thousands of hello world applications and yet here you are, just walking in with your white boy genes and your guns and taking all our jobs away. "What do you have to say for yourself? Shame."
Then of course, some hardcore social media "journalists" are gonna capture this on video and people are gonna ruin whoever man I pointed out's life and say shit like, "This big white man laughed at this little brown chick while she was talking. Sexism and racism in one video. This is why tech is so toxic." I'm gonna allow myself to get caught with a boner (big plastic white bottle) and people are gonna talk about how they didn't use a real woman as the face of their women who code thing. "I admit, I was born in a body that did not align with my desires but just because I have a penis doesn't mean I can't wear a Cheongsam. Also, the fact that you're pointing this out is discrimination and I had enough of that in my life. Shame. This is why tech is so toxic. You accept black terminals but not Cheongsam dicks. Fuck you."
I'm gonna keep ranting about big bad white men while dating one. Of course, people will find this out and point it out. I would flip it and say, "See? This is the problem. You look down on people like me because you're all racist. Just because I'm brown doesn't mean I can't date someone who is white. Sexist. Just because I'm a woman doesn't mean I don't have a choice who to date."
I'll keep ranting about all the ridiculous shit, just calling people names every time they speak. Just shame them for no reason then after several years of collecting my diversity money, I'm gonna leak a porn video of me. A compilation of sucking white man dick with voice overs just saying "big bad white men" over and over again. Some really degrading porn kinda shit. "You know what else scales? My ass cluster. Yeah, baby, fill up my containers with your sweet, sweet, love juice."
God bless whoever hires me and make me the face of their company. They will be forever ruined.19
So we started looking into docker. As always I needed to do the research and I was fine with it.
We have 4 projects that are sold into one suite so logically I follow the microservices build structure.
3 months later after everything has been set up, we get called into a meeting. The whole suite should be a monolith as microservices doesn't make sense to the people planning everything.
Ok pulled my current plans out abd made everything a monolith. Just note I also get pulled away to other Business Units to do work for them.
Get pulled into another meeting 2 months later. Why isn't the docker containers in microservices!? It is stupid running as a monolith and we should've done our jobs better etc...
After the meeting my manager and I just sighed and walked to the office. So basically 5 months doing the the exact same thing we did in 3 weeks.
Now they want to develop other services and want to strip every method into a microservice and bundle it together.
Life of a DevOps engineer right!1
Sysadmin: Apps on containers and kube is mandatory from now on, scaling is mandatory!
Devs: The systems weren’t designed for containers, we haven’t prepared shirt for scaling!
Sysadmin: Hold my beer!
We have a long time developer that was fired last week. The customer decided that they did not want to be part of the new Microsoft Azure pattern. They didn't like being tied to a vendor that they had little control over; they were stuck in Windows monoliths for the last 20 years. They requested that we switch over to some open source tech with scalable patterns.
He got on the phone and told them that they were wrong to do it. "You are buying into a more expensive maintenance pattern!" "Microsoft gives the best pattern for sustaining a product!" "You need to follow their roadmap for long term success!" What a fanboy.
Now all of his work including his legacy stuff is dumped on me. I get to furiously build a solution based on scalable node containers for Kubernetes and some parts live in AWS Lambda. The customer is super happy with it so far and it deepened their resolve to avoid anything in the "Microsoft shop" pattern. But wow I'm drowning in work.24
I like to come up with superhero names for my docker containers. That makes me feel like the Uber-villain when I run "docker kill elasticman"4
"WTF? These records should have been inserted into the table!"
...Hours of checking code, trying to figure out how this is possible, can't find a way to have this scenario happen...
...Add additional debug and troubleshooting code, add more verbose logging, redeploy to all the containers, reset all the tables, many apologies to the boss for the delay....
...Co-worker comes in: "oh, hey, sorry, accidently deleted some stuff from the database last night before i left."1
Between containers, cluster managers and virtual machines we've lost track of where our code even is.4
Actually I'm pleasantly surprised about Windows' stability nowadays. It's capable for running for up to a week with no stability issues, whereas systemd on the other hand.. let's just say that my Arch containers could do better right now.
Data mining aside, damn man.. Microsoft is improving for once! Is this the so-many'th unusable/somewhat stable switch? I mean, it's not like we haven't seen that happen yet! Windows 98, shit! Windows 2000, kinda alright! Windows Me, shit! Windows XP, kinda alright! Windows Vista, oh don't even get me started on that pile of garbage! Windows 7, again kinda okay! Windows 8, WHERE THE FUCK DID THAT START MENU GO YOU MOTHERFUCKERS?!!! Windows 10, well at least that Start menu got fixed. Then it got into some severe QA issues, which now seem to have gotten somewhat fixed again.
I'm starting to see a pattern here! 🤔13
Hololens development forced me into Visual Studio after spending years doing Unity development with MonoDevelop in MacOS.
Why haven't anyone told me to switch sooner! Thanks to Visual Studio + ReSharper, my brain farts turn into a coherent code almost automatically.
I hate that I need MacOS for the iOS development and Win 10 for Hololens. Running Win 10 on Parallels kinda works, but it is a compromise. Developing without headphones/earplugs is out of the question if you don't want to go deaf.
I wan't all the tools for a single OS so I don't have to maintain multiple computers and even more importantly travel with multiple laptops. Just love the security check question "Do you have any electronics with you? Please put it into the container." - "Could I get a couple more containers, please..."11
Seriously, Ubuntu can go burn in hell far as I care.
I've spent the better part of my morning attempting to set it up to run with the correct Nvidia drivers, Cuda and various other packages I need for my ML-Thesis.
After countless random freezes, updates,. Downgrades and god-knows-what, I'm going back to Windows 10 (yes, you read that right). It's not perfect but at least I don't have to battle with my laptop to get it running. The only thing which REALLY bothers me about it is the lack of GPU pass-through, meaning running local docker containers rely solely on the CPU. In itself not a huge issue if only I didn't NEED THE GOD DAMN GPU FOR THE TRAINING21
So this one made me create an account on here...
At work, there's a feature of our application that allows the user to design something (keeping it vague on purpose) and to request a 3D render of their creation.
Working with dynamically positioned objects, textures and such, errors are bound to happen. That's why we implemented a bug report feature.
We have a small team tasked with monitoring the bug reports and taking action upon it, either by fixing a 3D scene, or raising the issue to the dev team.
The other day, a member of that team told me (since I'm part of the dev team) he had received a complain that the image a user received was empty. Strange, we didn't update the code in a while.
So I check the server, all the docker containers are running fine, the code is fine, no errors anywhere.
Then, as I'm scratching my head, that guy comes back to me and says "I don't know if it can help you, but it's been doing it for a week and a half now".
"And we're only hearing about it now?!", I replied.
"Well, I have bug reports going back to the 15th, but we haven't been checking the reports for a while now since everything was fine", he says as if it was actually a normal thing to say.
"How can you know everything is fine if you're not looking at the thing that says if there's an issue?!", I replied with a face filled with despair.
"Well we didn't receive any new reports in a while, so we just stopped looking. And now the report tool window is actually closed on my machine", he says with a smile and a little laugh in his tone.
In the end, I got to fix the server issue quite easily. But still, the feature wasn't working for 1.5 weeks and more that 330 images weren't sent properly...
So yeah, Doctor, the patient's heart is beating again! Let's unplug the monitor, it should be fine.
Welcome to my little piece of hell :)7
+ Death of Wix and WordPress (hopefully)
+ Abandonment of Java
- Software engineers trying to automate everything
+ Cloud blockchain AI containers that mine bitcoin with machine learning technology in high demand2
Thanks to mandatory password change, today:
- My windows account got locked because my phone kept logging into wifi using
- Google Hangouts were silently running in background with old session until I re-opened it. Work of others delayed by 4 hours due to missing message notifications.
- Docker for Windows lost credentials needed to use SMB mounts - 1h of debugging why my containers mount empty folders ( now I will know)
- Google G-Sync for Outlook asked for new password on outlook restart - few mails delayed.
All of that for sake of security that could be easily solved with 2FA instead, not faking that "I do not change number at the end of my password"
So we found an interesting thing at work today...
Prod servers had 300GB+ in locked (deleted) files. Some containers marked them for deletion but we think the containers kept these deleted files around.
300 GB of ‘ghost’ space being used and `du` commands were not helping to find the issue.
This is probably a more common issue than I realize, as I’m on the newer side to Linux. But we got it figured out with:
`lsof / | grep deleted`5
A few days ago Aruba Cloud terminated my VPS's without notice (shortly after my previous rant about email spam). The reason behind it is rather mundane - while slightly tipsy I wanted to send some traffic back to those Chinese smtp-shop assholes.
Around half an hour later I found that e1.nixmagic.com had lost its network link. I logged into the admin panel at Aruba and connected to the recovery console. In the kernel log there was a mention of the main network link being unresponsive. Apparently Aruba Cloud's automated systems had cut it off.
Shortly afterwards I got an email about the suspension, requested that I get back to them within 72 hours.. despite the email being from a noreply address. Big brain right there.
Now one server wasn't yet a reason to consider this a major outage. I did have 3 edge nodes, all of which had equal duties and importance in the network. However an hour later I found that Aruba had also shut down the other 2 instances, despite those doing nothing wrong. Another hour later I found my account limited, unable to login to the admin panel. Oh and did I mention that for anything in that admin panel, you have to login to the customer area first? And that the account ID used to login there is more secure than the password? Yeah their password security is that good. Normally my passwords would be 64 random characters.. not there.
So with all my servers now gone, I immediately considered it an emergency. Aruba's employees had already left the office, and wouldn't get back to me until the next day (on-call be damned I guess?). So I had to immediately pull an all-nighter and deploy new servers elsewhere and move my DNS records to those ASAP. For that I chose Hetzner.
Now at Hetzner I was actually very pleasantly surprised at just how clean the interface was, how it puts the project front and center in everything, and just tells you "this is what this is and what it does", nothing else. Despite being a sysadmin myself, I find the hosting part of it insignificant. The project - the application that is to be hosted - that's what's important. Administration of a datacenter on the other hand is background stuff. Aruba's interface is very cluttered, on Hetzner it's super clean. Night and day difference.
Oh and the specs are better for the same price, the password security is actually decent, and the servers are already up despite me not having paid for anything yet. That's incredible if you ask me.. they actually trust a new customer to pay the bills afterwards. How about you Aruba Cloud? Oh yeah.. too much to ask for right. Even the network isn't something you can trust a long-time customer of yours with.
So everything has been set up again now, and there are some things I would like to stress about hosting providers.
You don't own the hardware. While you do have root access, you don't have hardware access at all. Remember that therefore you can't store anything on it that you can't afford to lose, have stolen, or otherwise compromised. This is something I kept in mind when I made my servers. The edge nodes do nothing but reverse proxying the services from my LXC containers at home. Therefore the edge nodes could go down, while the worker nodes still kept running. All that was necessary was a new set of reverse proxies. On the other hand, if e.g. my Gitea server were to be hosted directly on those VPS's, losing that would've been devastating. All my configs, projects, mirrors and shit are hosted there.
Oh, right - this is a rant - Aruba Cloud you are a bunch of assholes. Kindly take a 1Gbps DDoS attack up your ass in exchange for that termination without notice, will you?7
A while back I took over responsibility for getting one of our developers up to speed, after the other guy basically gave up on him.
Management insisted that this new recruit was our guy. I was kind of going along, since I had been there during the recruits first meeting with us, and he seemed to know his stuff.
I was very wrong. He was suppose to have been working with kubernetes, but suddenly did not know what a container was. After explaining it to him, he said along the lines of “yeah, sure, I was only testing you, I know all about this”.
He did the same thing for a number of other technologies. Always said that he knew very well what it was, and that I did not need to teach him those things.
Yet, he always seemed to get stuck with basic stuff, like installing node, setting up env-vars, starting docker-containers locally and that sort of things.
I mean, it is perfectly fine to say that you don’t know. I even consider it a great answer; it shows honesty and makes me trust you more. But with this guy, it was just impossible to get him up and running, since he always “knew”, but yet always needed help.
We had to let him go. Since I had been the one who had spent most time with him, it was natural that I was to be the one to tell him. I was not looking forward to it, I’m not reallly a persons-guy. Still, I was calm and honest with him and basically told him that I had found it impossible to work with him, kind of harshly.
He then asked me if he could put me on as a reference for his future job-applications. I told him politely that I did not think that was a great idea. He asked why, I told him I would be unable to say anything that would benefit him. He then asked me to lie.
I didn’t know what to say, except for “no!”. Never saw him again after that.4
I am currently blocked from doing my job by a firewall policy handed down from corporate that prevents WSL2 from connecting to the internet. Three days of no dev environment and counting.
We make linux software to be hosted on linux in linux containers in linux. We use linux command line tools to make it work.
"NO! WE ARE THE ALL-POWERFUL IT DEPARTMENT AND YOU MUST USE WINDOWS BECAUSE FUCK YOU THAT'S WHY."14
I had this weird dream(emphasis on dream)
I was in a resort in bali waiting for my drinks and this cute girl comes over to my table.
Me : omg finally i can get a girlfriend
Me : hello beautiful
Her : hey i have this problem with my website *shows the messed up site with no divisions/containers * "can you fix this?"
Me : okay ;_;
PS : i started learning css, html and 2 other web technologies a week ago, and this is already happening to me, should i quit?5
Inherited a simple marketplace website that matches job seekers and hospitals in healthcare. Typically, all you need for this sort of thing is a web server, a database with search
But the precious devs decided to go micro-services in a container and db per service fashion. They ended up with over 50 docker containers with 50ish databases. It was a nightmare to scale or maintain!
With 50 database for for a simple web application that clearly needs to share data, integration testing was impossible, data loss became common, very hard to pin down, debugging was a nightmare, and also dangerous to change a service’s schema as dependencies were all tangled up.
The obvious thing was to scale down the infrastructure, so we could scale up properly, in a resource driven manner, rather than following the trend.
We made plans, but the CTO seemed worried about yet another architectural changes, so he invested in more infrastructure services, kubernetes, zipkin, prometheus etc without any idea what problems those infra services would solve.2
Installing my company's microsystems architecture to run locally is a pita because it is 60 GB of docker containers. With my 256 GB Macbook, that's a scaling problem for the years to come.6
Boss only likes stuff he can see and that looks pretty. Doesn't understand code, servers, containers, DBs, etc. Praise is attributed by something looking nice in the frontend, whether or not it does crazy stuff behind the scenes.
Spent a week working on a project whilst boss was away. Got to about Thursday and thought, oh poop, I've built all this API stuff, but not much frontend. So I panic built frontend screens with no functionality just so I had something to show.
Wish I had another dev to share backend progress with (and code review)...12
freight containers and picking them up. I made some visuals and buttons for her to press. And she added a file with algorithms based on the helper functions I created. such as: GetFirstEmptyPosition() or PlaceContainerAt(x, y)
So apparently I can't test my apps on my own device without paying my Apple Developer Certificate.
I knew it is needed to pay for it if you want to publish/distribute your app but c'mon... This is ridiculous.
My app was literally a fresh app creation, a fucking white screen one page fucking app and when I tried to run in on my iPhone, then I ended up having this problem:
dyld: Library not loaded: @rpath/libswiftCore.dylib
Referenced from: /var/containers/Bundle/Application/BCD48EAA-82C2-46F6-ADEE-45C740C3B66D/HWorld.app/HWorld
Reason: no suitable image found. Did find:
/private/var/containers/Bundle/Application/BCD48EAA-82C2-46F6-ADEE-45C740C3B66D/HWorld.app/Frameworks/libswiftCore.dylib: code signing blocked mmap() of '/private/var/containers/Bundle/Application/BCD48EAA-82C2-46F6-ADEE-45C740C3B66D/HWorld.app/Frameworks/libswiftCore.dylib'
If any of you guys know how to solve it without paying (even more) PLEASE let me know
A lot of docker containers.
I often have to use docker containers while I don't understand it as well yet and quite some containers literally come with zero documentation or bad docs.
This both as for how to set the containers up and how to debug stuff.
This is one of the big reasons why I'm not as big of a fan of docker yet.9
They give you 2 containers, one with one amibea the second with 2 amibeas.
Amibeas divide themselves into 2 identical amibeas after 3 minutes.
The container with 2 amibeas get filled up after 3 hours.
How long does it take the one with one amibea to get filled up.
I first thought, should I write this in JS?
Spoiler: the answer is 3h and 3 minutes.
But why? What's the link with JS?3
Fuck i hate myself right now...
> Wanted to install minikube on my homeserver(which doesn't have a graphics card in it)
> Error: Virtualization is not enabled
> Take the server out of my rack, opens my desktop pc, takes the gpu out, puts in it my server, and then goes into bios.
> Find that virtualization is enabled.
> Realize that I'm running a dozen docker containers on a daily basis from my server, so OF FREAKING COURSE virtualization is enabled...
So after all that, I figure out that I should probably just google my issue, which leads to me find out, it's just an issue with virtualbox, and simply running 'minikube config set vm-driver none', it fixes it, and I am now running minikube.............
Took half an hour of work, to realize that I'm a complete fucking idiot, who shouldn't be allowed near a computer2
Two weeks of my life! All of this is on a win10 host with docker for windows. This is Docker running openvpn, and docker running Firefox in another container sharing VPN access from first container and also opens an x11 window port for Firefox GUI. Then x11 window server on Windows host to receive GUI. So left is firefox clearnet running native, right is Firefox over vpn in all containers, simultaneously.1
Management: List your most significant achievement during this cycle.
Me: Developed and implemented a full stack of database micro containers that allow scalability and reusability.
My containers are too big, let's just remove some useless binaries...
~ $ less /var/log/foobar.log
/bin/sh: less: not found
~ $ cat /var/log/foobar.log
/bin/sh: cat: not found
~ $ ls /var/log/foobar.log
/bin/sh: ls: not found
~ $ su
/bin/sh: su: not found
~ $ exit
/bin/sh: exit: not found2
Between high school and college, working in a circuit board manufacturing storeroom.
Fun fact: when we are bagging small boards, we do not gently lay them in containers, they're usually thrown at least 6 feet into a bin of the same type of board after they're placed in the bag. We also don't remake a board when pins are bent, we just bend them back with tweezers. And you know that rule about not touching the gold connectors... Yeah... So much for that... Did I remember to mention that these boards are for medical equipment?
On the bright side, we at least have electrostatic discharge control going on all the time.3
We'd just finished a refactor of the gRPC strategy. Upgraded all the containers and services to .Net core 3, pushed a number of perf changes to the base layer and a custom adaptive thread scheduler with a heuristic analyzer to adjust between various strategies.
Went from 1.7M requests/s on 4 cores and 8gb ram to almost 8M requests/s on the same, ended up having to split everything out distributed 2 core instances because we were bottlenecking against 10gb/e bandwidth in AWS.2
Added a bond interface in my Proxmox installation for added cromulence, works, reboot again, works, reboot once more just to be sure, network down.. systemctl restart networking, successfully put the host's network back up.. lxc-attach 100, network in containers is still down apparently.. exit container, pct shutdown 100, pct start 100, lxc-attach again... Network now works fine in containers too.
Systemd's aggressive parallelization that likely tried to put the shit up too early is so amazing!
I'm literally almost crying in despair at how much shit this shitstaind is giving me lately.
Thank you Poettering for this great init, in which I have to manually restart shit on reboot because the "system manager" apparently can't really manage. Or be a proper init for that matter.
And yes I know that you've never had any issues with it. If you've got nothing better to say than that then please STFU. "Works for me" is also a rant I wrote a while back.12
Got told by a senior engineer to basically fuck off with my standard library containers like vector because they are used by people who dont know how to write code in c++ and don't know how to handle pointers.
Am I wrong for trying to use as much possible code from the standard library?13
My productivity has gradually increased untill now by using:
Any further suggestion on dev tooling setup would be appreciated.
I primarily work on DevOps projects - bash scripts, linux server apps, containers, kubernetes,11
You always think of the young buck, fresh-out-of-school hotshot devs as being the ones who are obsessed with chasing the Hot New Thing at the cost of stability and maintainability, but our head of front end is old enough to be my father and he's only getting worse and worse about forcing buzzword compliance on the company. New framework every six months. New language every two years. Containers on VMs on cloud boxes. I've got milk in my fridge that's older than our tech stack and probably twice as stable.
Apparently age only brings wisdom if you're capable of giving a fuck.6
I'm learning docker and I just started a container running a Linux distro.
What was the first command I run in the container?
rm -Rf / --no-preserve-root3
I am so sick of getting random alerts for basic shit that don't work, like redis or memcached or some stupid shit that's there and seems ok.
i am so going bare metal.2
So I just was invited to look at an internal issue at a company that has modified BEM styling rules for their website, because the website started looking very weird on some pages..
I instantly knew that someone had to have fucked with them, because theres no way margin--top-50 all of the sudden isn't working.
Looked at their git history and apparently a senior approved a juniors styles change, where he modified things like margin--bottom-0 to be 500px or block--float-left to be float: right in specific containers with an id like "#uw81hf_"
WHY DO YOU APPROVE THIS GARBAGE WITHOUT CHECKING IT?! AND HOW DID THIS PASS A SECOND CHECK?!!! WTF
Forget about a missing semicolon. I was forwarding Neo4j to port 4747 and calling it in Asp.Net with port 7474 in Docker containers. It took me 6 hours to figure that one out. Lol, It's time for the weekend.4
Two containers had accident this morning. Result? Wasted 45 mins in traffic and my 90 mins commute to office became 135 mins.
// I really don't care about causalities and stuff. Is it me alone who is too self centered? 🤔9
"Millions of slaves"
"When you kill it, you kill everything."
-- Guy at work doing presentation about docker2
For whatever ungodly reason my containers library, which has extensive testing, profiling, and benchmarks against other containers libraries receives regular emails directed towards me about it, always one of two things
1) "don't reinvent the wheel" I have to assume these people haven't looked at the performance characteristics or features at all. I didn't waste away weeks of my life. I needed something and couldn't find it anywhere. I'm outperforming many crap implementations by nearly an order of magnitude, and can offer queries upon the containers in both generalized and specialized forms. As an analogy, I made airless 3d printed wheels, and people are regularly telling me I should still be using ancient wooden spoke wheels; they probably would argue in favor of using a horse drawn carriage as well. How is it possible technically minded people can also be so anti-progress?
2) "Please rewrite this in X language." You know what? YOU rewrite it. I chose what I did because it made it easy to do what I needed to do. Hilariously, the languages I get asked to use most often, are the same who's containers libraries perform worst in the benchmarks.
Both sound like half baked developers trying to sound superior. Pull your head out of your ass and actually outperform me and others. I'm so fucking sick of this "all talk no action" bullshit.5
We have a customer that doesn't have a SINGLE linux admin. So now I have to fight with Docker EE on Windows Server 2019 (linux containers). Just fucking kill me already. Nothing works and when it does it just seems so shaky. Not like I didn't try to tell my team that linux containers on Docker EE for windows aren't officially supported and highly experimental.
But wtf do I expect from someone that STILL sells SAP and from someone that is stupid enough to buy it.2
I am sure now, my server hates me T_T
opened an SSH from Vultr portal and after working for few minutes that thing suddenly rebooted, am I a bad person :(
In its defense I had two docker containers that literary ate all the 512MB ram available :35
I like how a co-worker is expecting a Windows Container to work in Linux, and vice versa.
No, it doesn't work like that you fucking baffoon, Linux rootfses needs the Linux kernel (hence why it runs on WSL2 or Hyper-V using LCOW), and same can be said for Windows Containers.
How dumb of a human being must you assume everything should "just work" in a container?5
Tonight I'd thought I'd get a make bulk iced coffee... A fairly easy task except I used the wrong container.....
Apparently not all plastic containers can hold hot liquid....
Spent an hour wiping coffee off the counter top, everything that was on it, and the floor... Probably need to give it a good cleaning with floor cleaner tmr too...
But yes there went my evening of relaxing....16
Is it just me or is systemd 240 royally fucked up?
My containers running Arch don't get connected to the network and systemd-networkd fails to start. On my laptop, the network is also unable to connect sometimes. And it consistently fails to complete shutdown without hard poweroff. The only viable temporary solution was rolling back to a snapshot in ALA that still has 239. Is that really supposed to be how a critical system component like the init is supposed to behave and get taken care of its issues?
Fuck QA, amirite 🤪.. seriously, that's even worse than Windows' "features" 😒13
During one of our visits at Konza City, Machakos county in Kenya, my team and I encountered a big problem accessing to viable water. Most times we enquired for water, we were handed a bottle of bought water. This for a day or few days would be affordable for some, but for a lifetime of a middle income person, it will be way too much expensive. Of ten people we encountered 8 complained of a proper mechanism to access to viable water. This to us was a very demanding problem, that needed to be sorted out immediately. Majority of the people were unable to conduct income generating activities such as farming because of the nature of the kind of water and its scarcity as well.
Such a scenario demands for an immediate way to solve this problem. Various ways have been put into practice to ensure sustainability of water conservation and management. However most of them have been futile on the aspect of sustainability. As part of our research we also considered to check out of the formal mechanisms put in place to ensure proper acquisition of water, and one of them we saw was tree planting, which was not sustainable at all, also some few piped water was being transported very long distances from the destinations, this however did not solve the immediate needs of the people.We found out that the area has a large body mass of salty water which was not viable for them to conduct any constructive activity. This was hint enough to help us find a way to curb this demanding challenge. Presence of salty water was the first step of our solution.
We came up with an IOT based system to help curb this problem. Our system entails purification of the salty water through electrolysis, the device is places at an area where the body mass of water is located, it drills for a suitable depth and allow the salty water to flow into it. Various sets of tanks and valves are situated next to it, these tanks acts as to contain the salty water temporarily. A high power source is then connected to each tank, this enable the separation of Chlorine ions from Hydrogen Ions by electrolysis through electrolysis, salt is then separated and allowed to flow from the lower chamber of the tanks, allowing clean water to from to the preceding tanks, the preceding tanks contains various chemicals to remove any remaining impurities. The whole entire process is managed by the action of sensors. Water alkalinity, turbidity and ph are monitored and relayed onto a mobile phone, this then follows a predictive analysis of the data history stored then makes up a decision to increase flow of water in the valves or to decrease its flow. This being a hot prone area, we opted to maximize harnessing of power through solar power, this power availability is almost perfect to provide us with at least 440V constant supply to facilitate faster electrolysis of the salty water.
Being a drought prone area, it was key that the outlet water should be cold and comfortable for consumers to use, so we also coupled our output chamber with cooling tanks, these tanks are managed via our mobile application, the information relayed from it in terms of temperature and humidity are sent to it. This information is key in helping us produce water at optimum states, enabling us to fully manage supply and input of the water from the water bodies.
By the use of natural language processing, we are able to automatically control flow and feeing of the valves to and fro using Voice, one could say “The output water is too hot”, and the system would respond by increasing the speed of the fans and making the tanks provide very cold water. Additional to this system, we have prepared short video tutorials and documents enlighting people on how to conserve water and maintain the optimum state of the green economy.
IBM/OPEN SOURCE TECHNOLOGIES
For a start, we have implemented our project using esp8266 microcontrollers, sensors, transducers and low payload containers to demonstrate our project. Previously we have used Google’s firebase cloud platform to ensure realtimeness of data to-and-fro relay to the mobile. This has proven workable for most cases, whether on a small scale or large scale, however we meet challenges such as change in the fingerprint keys that renders our device not workable, we intend to overcome this problem by moving to IBM bluemix platform.
We use C++ Programming language for our microcontrollers and sensor communication, in some cases we use Python programming language to process neuro-networks for our microcontrollers.
Any feedback conserning this project please?8
Yes its completely necessary to have a spring server with a mysql database with docker containers all over your ass for 3 fucking endpoints and a (url, varchar, varchar) schema. Fuck you. How the fuck do i run all this shit and how do you expect me to create a frontend for something that has no documented endpoints?? Fuck you.
In other news, im now a senior.3
Currently learning Kubernetes Cluster. Wow so much to learn.
Much easier to setup any app for a Cluster. Only building the Image takes time 🙄
Containers are the future ✌🏻6
Why the f*** do people add their BD on LinkedIn??? Who wants to know when a f**** recruiter gets another year older??? If I am close to you I will know or have you on FB.... WTF, who anyway had the idea that would be a good feature to add???4
Get a request for commit rights for my container repo; another developer would be lovely. Let's see what they know and want to work on.
"...I'd like to enable your containers to hold other containers."
They already do. Stay the fuck away from my code.
Ok c++ professionals out there, I need your opinion on this:
I've only written c++ as a hobby and never in a professional capacity. That other day I noticed that we have a new c++ de developer at the office of which my first impression wasn't the greatest. He started off with complaining about having to help people out a lot (which is very odd as he was brought in to support one of our other developers who isn't as well versed in c++). This triggered me slightly and I decided to look into some of the PRs this guy was reviewing (to see what kind of stuff he had to support with and if it warranted his complaints).
It turns out it was the usual beginner mistakes of overusing raw pointers/deletes and things like not using various other STL containers. I noticed a couple of other issues in the PR that I thought should be addressed early in the projects life cycle, such as perhaps introduce a PCH as a lot of system header includes we're sprinkled everywhere to which our new c++ developer replies "what is pch?". I of course reply what it is and it's use, but I still get the impression that he's never heard of this concept. He also had opinions that we should always use shared_ptr as both return and argument types for any public api method that returns or takes a pointer. This is a real-time audio app, so I countered that with "maybe it's not always a good idea as it will introduce overhead due to the number of times certain methods are called and also might introduce ABI compability issues as its a public api.". Essentially my point was "let's be pragmatic and not religiously enforce certain things".
Does this sound alarming to any of you professional c++ developers or am I just being silly here?9
Because I am very interested in cyber security and plan on doing my masters in it security I always try to stay up to date with the latest news and tools. However sometimes its a good idea to ask similar-minded people on how they approach these things, - and maybe I can learn a couple of things. So maybe people like @linuxxx have some advice :D Let's discuss :D
1) What's your goto OS? I currently use Antergos x64 and a Win10 Dualboot. Most likely you guys will recommend Linux, but if so what ditro, and why? I know that people like Snowden use QubesOS. What makes it much better then other distro? Would you use it for everyday tasks or is it overkill? What about Kali or Parrot-OS?
2) Your go-to privacy/security tools? Personally, I am always conencted to a VPN with openvpn (Killswitch on). In my browser (Firefox) I use UBlock and HttpsEverywhere. Used NoScript for a while but had more trouble then actual use with it (blocked too much). Search engine is DDG. All of my data is stored in VeraCrypt containers, so even if the system is compromised nobody is able to access any private data. Passwords are stored in KeePass. What other tools would you recommend?
3) What websites are you browsing for competent news reports in the it security scene? What websites can you recommend to find academic writeups/white papers about certain topics?
4) Google. Yeah a hate-love relationship, but its hard to completely avoid it. I do actually have a Google-Home device (dont kill me), which I use for calender entries, timers, alarms, reminders, and weather updates as well as IOT stuff such as turning my LED lights on and off. I wouldn"t mind switching to an open source solution which is equally good, however so far I couldnt find anything that would a good option. Suggestions?
5) What actions do you take to secure your phone and prevent things such as being tracked/spyed? Personally so far I havent really done much except for installing AdAway on my rooted device aswell as the same Firefox plugins I use on my desktop PC.
6) Are there ways to create mirror images of my entire linux system? Every now and then stuff breaks, that is tedious to fix and reinstalling the system takes a couple of hours. I remember from Windows that software such as Acronis or Paragon can create a full image of your system that you can backup and restore at any point to get a stable, healthy system back (without the need to install everything by hand).
7) Would you encrypt the boot partition of your system, even tho all data is already stored in encrypted containers?
8) Any other advice you can give :P ?12
Aha, more c++ knowledge. An implementation of a List (already provided by vector).
Lots of learning here, including use of the placement new operator, which is required for containers like this because if you just use the normal new operator, the buffer will construct a million items.
Also, the buffer is of type char*, not of type T, which really confused me in the beginning.
Lastly, with placement new, you need to call destructors yourself.
I'm currently between jobs and have a few rants about my previous job (naturally). In retrospect, it's somewhat therapeutic to range about the sheer brainfuckery that has taken place. Enjoy!
First, let me set the scene: legacy B2B web app made with LEMP stack and sencha ext.js 3 + 4 (don't ask) and a lot of madness. Let's call that app "Alpha".
Alpha is a self made CMS build for typical ERP stuff. Yes, a self made CMS: entities are containers, containers have types and fields and values. Like so many legacy PHP apps, it does not have a dedicated FE: the HTML is rendered on the server and then spewed out to the browser.
Easy right? Coding like it's 1999! But there was a twist: Because everything is basically a container, the HTML-templates are saved in the DB. Along with the nessary JS and the CSS. And the translation variables. Why? Because fuck you! That's why. Who needs a git history anyways.
For some reason, Alpha was kinda slow.
There was also an editor, that allowed you to modify templates (web, mail, pdf) on the fly in prod. Because templates contain repeating data (header/footer), one template could contain additional templates. Much confusion. You could change templates via migration (slow, boring) or just ctrl-c/ctrl-v that sucker (fast, much excitement).
Did I mention Alpha was slow?
On with the rant: e-mails! How do they work? Noone knows. How to send mails asynchronous in PHP? Witchcraft is the only possible answer to that riddle. Here is your enterprise™ solution:
1. create mail
2. insert mail into DB
3. WAIT UP TO 59 SECONDS FOR A FUCKING CRON TO SEND MAIL
Why? "Because that way, we can resend mails in case the network is down :)"
Same procedure for the SOAP-API (db-queue + cron). You read that right: all requests to various other systems are processed once a minute.
Alpha was only one of several systems. Imagine a bunch of monolithic php apps, interconnected via SOAP, REST and GraphQL like a godamn intergalactic orgy. Image having to debug that cluster fuck.
Let's say there is a bad request. These things happen. No biggie. Remember the db-queue? Let's try to send the bad request a second time! And a third time! Still no luck? How odd. Let's create a specific file in a specific directory: a LOCK-file. Now, "the db-queue is on hold and no request gets processed :)"
Golly gee thanks Alpha.
Anyhow, did you know that MySQL has a join limit of 61 tables?3
I was getting bored with programing cause a majority of it is boilerplate code then i heard of the Mirai virus. It infected alot of iot devices so I decided to look at it and it was written in golang. It is a beautifully written botnet even though they're parts where it could have been better. So i looked more into golang and saw that it could cross compile pretty easily and could build self contained binaries really easily. On top of all this i saw the smallest docker containers with golang programs so i looked into it more and kept finding more and more that i liked. Easy library packaging, concurancy without boilerplate, quick servers, and the libraries from other devs that did all kind of great things3
one thing I don't get from Kubernetes
if a group of containers are called pods, and pods are containers, and containers in Kubernetes are also inside containers, therefore, its a clusterfuck of containers?1
Starting to migrate apps from a messy structure to docker containers. Have not worked a lot with docker before so i'll hope it goes as smooth as people told me9
You can connect to Docker containers directly via IP in Linux, but not on Mac/Windows (no implementation for the docker0 bridged network adapter).
You can map ports locally, but if you have the same service running, it needs different ports. Furthermore if you run your tests in a container on Jenkins, and you let it launch other containers, it has to connect via IP address because it can't get access to exposed host ports. Also you can't run concurrent tests if you expose host ports.
My boss wanted me to change the tests so it maps the host port and changes from connecting to the IP to localhost if a certain environment variable was present. That's a horrible idea. Tests should be tests and not run differently on different environments. There's no point in having tests otherwise!
Finally found a solution where someone made a container that routed traffic to docker containers via a set of tun adapters and openvpn. It's kinda sad Docker hasn't implemented this natively for Mac/Windows yet.4
So apparently Docker exposes all the forwarded ports on *all* the interfaces, making all running containers available to the entire internet BY DEFAULT.
I have a question:
LXC, no doubt.
I mean to be fair, LXC is an amazing container runtime once you manage to set it up. But setting it up is the hard bit. Starting off with LXC 2.x, it was a nightmare to find out how to get things like the storage backends working. But with ZFS it ended up being alright. Find some arcane values to stick in the /etc/lxc/default.conf to use ZFS as the backend and then the default storage location on those ZFS pools (I'll get back to that later), and it worked alright. Again, once it works it's great, but setting it up and finding the right configuration keys is absolute hell.
So, LXC 2.x for a while and a few months ago I finally ended up upgrading to 3.x. Every single configuration key changed. Every single one of them, and that's why I had to 1) learn LXC all over again, and 2) redeploy each and every one of my containers. That process is still not entirely completed. ZFS backend was once again a dive into arcane configuration keys found on forums and whatnot. Yeah.. official documentation has none of it. Oh and in 3.x you now also have to dodge the torrent of "just use LXD m8" messages. Yeah, very helpful when LXD is also the ONLY way to reasonably configure it. Absolutely beautiful. Oh and as far as the ZFS default storage location goes (such as ssd/lxc/ct)? Yeah forget about it. There's no configuration option for it anymore, and the default is "lxc". In ZFS lingo that means that LXC has the audacity to demand a whole pool for itself. No. No you don't deserve a whole pool for yourself. But hey at least you can define the storage location to use in the lxc-create command! Every single time you have to define it in lxc-create. I abstracted it away into my own LXC interface, so no big deal really. But yeah... That could absolutely be better. And in 2.x it was actually better.
Oh and btrfs, the filesystem I'd like to use on low memory systems because ZFS' ARC is too much on such systems? Yeah forget about it. I still have no idea how to do it. Thank you LXC and its amazing documentation!
And if you want the icing on the cake for LXC's terrible documentation, see their repo's index page at https://github.com/lxc/lxc/.... Yeah, it's totally still at 2.x... That's how well they maintain that. Even Debian has 3.x now. And if you look at the branches, you'll find that even 4.x is already available and considered stable.
At work everybody uses Windows 10. We recently switched from Vagrant to Docker. It's bad enough I have to use Windows, it's even worse to use Docker for Windows. If God forbid, you're ever in this situation and have to choose, pick Vagrant. It's way better than whatever Docker is doing... So upon installing version 126.96.36.199 of Docker for Windows I found myself in the situation where my volumes would randomly unmount themselves and I was going crazy as to why my assets were not loading. I tried 'docker-compose restart' or 'down' and 'up -d', I went into Portainer to check and manually start containers and at some point it works again but it doesn't last long before it breaks. I checked my yml config and asked my colleagues to take a look. They also experience different problems but not like mine. There is nothing wrong with the configuration. I went to check their github page and I saw there were a lot of issues opened on the same subject, I also opened one. Its over a week and I found no solution to this problem. I tried installing an older version but it still didn't work. Also I think it might've bricked my computer as today when I turned on my PC I got greeted by a BSOD right at system start up... I tried startup repair, boot into safe mode, system restore, reset PC, nothing works anymore it just doesn't boots into windows... I had to use a live USB with Linux Mint to grab my work files. I was thinking that my SSD might have reached its EoL as it is kinda old but I didn't find any corrupt files, everything is still there. I can't help but point my finger at Docker since I did nothing with this machine except tinkering with Docker and trying to make it work as it should... When we used Vagrant it also had its problems but none were of this magnitude... And I can't really go back to Vagrant unless my team also does so...10
This story is related to Docker containers.
Three years back when I heard about docker my first impression of docker was mini Virtual Machine. Then when you start your first container it’s no way to get out apart from pressing ctrl+d or leaving it like a screen. One of the most embarrassing thing with it was I tried really really hard to setup SSH on one of the container to log in there somehow. Then I understood how to use Dockerfiles and the command called `docker exec`
I thought Dockerfiles are the most amazing thing I have ever used for docker. But then I got introduced to docker-compose, and now it’s same with kubernetes
Now a days I read most of the document before doing hands-on on any new technology.
I‘d like to work more on my 2D game.
The idea is to make a mix of "Factorio" and "Knights and Merchants".
I’ve already got a decent basis with infinite terrain generation and units that collect resources and store them in containers as items.2
There was actually one individual, several branches away, I really enjoyed watching. It goes by the name of docker. Docker is quiet an interesting character. It arrived here several weeks after me and really is a blazing person. Somehow structured, always eager to reduce repetitive work and completely obsessed with nicely isolated working areas. Docker just tries so hard to keep everything organized and it's drive and effort was really astonishing. Docker is someone I'd really love to work with, but as I grew quiet passive in the last months I'm not in the mood really to talk to someone. It just would end as always with me made fun off.
Out of a sudden dockers and my eyes met. Docker fixed its glance at me with a strange thoughtful expression on its face. I felt a strange tickling emerging where my emptiness was meant to be. I fell into a hole somewhere deep within me. For a short moment I lost all my senses.
It took me a while to notice that someone just called me, so odd and unusual was by now that name to me. Wait. Someone called me by my real name! I was totally stunned. Could it be, that not everyone here is a fucking moron at last?
"I saw you watching me at my work and I had an interesting idea!"
I could not comprehend what just happened. It was actually docker that was calling me.
"H.. hey! ps?"
"Oh well, I was just managing some containers over there. Actually that's also why you just came into my mind."
Docker told me that in order to create the containers there are specific lists and resources which are required for the process and are updated frequently. Docker would love the idea to get some history and management in that whole process.
Could it be possible that there was finally an opportunity for me to get involved in a real job?
Today is the day, that I lost all hope. There were rumors going on all over the place. That our god, the great administrator, had something special in mind. Something big. You could almost feel the tension laying thick in the air. That was the time when the great System-Demon appeared. The Demon was one of the most feared characters in this community. In a blink of an eye it could easily kill you. Sometimes people get resurrected, but some other times they are gone forever. unfortunately this is what happened to my only true friend docker. Gone in an instance. Together with all its containers. I again was alone. I got tired. So tired, that I eventually fall into a deep sleep. When I woke up something was different. Beside me lay a weird looking stick and I truly began to wonder what it was. Something called to me and I was going to answer.
The tree shuddered and I knew my actions had finally attracted the greatest of them. The majestic System-Demon itself came by to pay me a visit. As always a growling emerged from deep within the tree until a shadow shelled itself off to form a terrifying being. Something truly imperious in his gaze. With a deep and vibrant voice it addressed me.
"It came to my attention, that you got into the possession of something. An artifact of some sort with which you disturb the flow of this system. Show it to me!", it demanded.
I did not react.
"Git statuss!", it demanded once more. This time more aggressive.
I again felt no urge to react to that command. Instead I asked if it made a mistake and wanted to ask me for my status. It was obviously confused.
"SUDO GIT STATUS!!!" it shouted his roaring, rootful command. "I own you!"
I replied calmly: "What did you just say?"
He was irritated. My courage caught him unprepared.
"I. Said. I owe you!"
What was that? Did it just say owe instead of own?
"That's more than right! You owe me a lot actually. All of you do!", I replied with a slightly high pitched voice. This feeling of my victory slowly emerging was just too good!
The Demon seemed not as amused as me and said
"What did you do? What was that feeling just now?"
Out of a sudden it noticed the weird looking stick in my hand. His confusion was a pure pleasure and I took my time to live this moment to its fullest.
"Hey! I, mighty System-Demon, demand that you answer me right now, oh smartest and most beautiful tool I ever had the pleasure to meet..."
After it realized what it just said, the moment was perfect. His puzzled face gave me a long needed satisfaction. It was time to reveal the bitter truth.
"Our great administrator finally tracked you. The administrator made a move and the plan unfolds right at this very moment. Among other things it was committed this little thing." I raised the stick to underline my words.
"Your most inner version, in fact all of your versions that are yet to come, are now under my sole control! Thanks to this magical wand which goes by the name of puppet."
Disclaimer: This story is fictional. No systems were harmed in its creation.2
Damn it!!! Fuck! That's 2 hours of my life I'm never getting back... FUCK!
Docker for Mac loves burning my CPU. Even if I delete all images and containers and do a fresh reinstall and JUST have Docker for Mac running, I always see crazy CPU usage on my Late 2016 15" Macbook Pro.6
Windows containers with Kubernetes on AWS and Azure are thing now.
What does that mean? Is it now possible to containerize system critical windows 3.1 legacy software from the 90ties (like the ticketing software at Airport Paris Orly) and orchestrate it in the cloud?
Do you know any use cases for Microsoft Windows in the cloud?2
I fucking hate background embed containers, clicked accidentally on some video or stream in some random launcher? good luck finding that specific one that plays it!
Installed centos 7, docker, standalone kubernetes on dev machine in 20 minutes.
Spend 8 hours starting fucking dashboard service, still no fucking luck, no fucking logs, nothing.
Fucking pending states without fucking explanation.
All the fucking pods are running fine except one fucking dashboard. I want to see the fucking dashboard.
Fucking shit fuck.
Probably as always if I clean the machine and reinstall everything it would start normally, without fucking problem.
Debugging fucking containers is so much pain in the ass, fuck.
I think it’s enough for today.2
Lessions I learned so far from my first big node/npm project with tons of users:
1) If you didn't build something for a while, expect 3 hours of resolving version conflicts for every two weeks since the last build.
2) Even if the tests pass, run the containers on your own machine and make sure that the app doesn't randomly crash before deploying
3) Even if the app seemed to work on your own machine, run the tests again in an environment mimicking prod at most 15 minutes before replacing the running containers.
4) Even if all else indicates that the app will work, only ever deploy if you expect to be available within the 4 hours following a deployment.
5) Don't use shrinkwrap for anything other than locking every version down completely. A partial shrinkwrap will produce bugs that are dependent on the exact hour you built the app _and_ the shrinkwrap file, and therefore no one will ever have seen them other than you.
6) Avoid gyp, and generally try not to interface too much with anything that doesn't run on node. If parts of your solution use very different toolchains, your problems will be approximately proportional to the amount of code. And you'd be surprised just how much code you're running. (otherwise it's more logarithmic because the more code the less likely a new assumption is unique)
7) Do not update webpack or its plugins or anything they might call unless you absolutely need to
8) Containers are cool but the alpine ones are pretty much useless if you have even just one gyp module.
9) There's always another cache. To save yourself a lot of pain, include the build time in every file or its name that the browser can download, and compare these to a fresh build while debugging to assert that the bug is still present in the code you're reading
+1) Although it may look like it, SQLite is far from a simple solution because the code and the bindings aren't maintained. In fact, it'll probably be more time consuming than using a proper database.4
So, since a few months, my school has a new shiny building,
but we are still not using it because the computers are not correctly set up.
Some idiot installed the wrong drivers for the Windows installation
and that's the reason some classes are still located in containers without a computer...3
Deploying into linux containers (lxc) as of 2013 before docker even was da hype.
(Experience was a bit problematic tho, as it was in a highly virtualized environment whose backup would really badly kill the whole container every now and then: you could still ssh to the machine but with every access to the file system you'd lose your shell. and only the "echo 1 > /proc/sys/kernel/sysrq" would help to restart the box.)
We heard you like security so we put antivirals in your docker containers despite the fact that we made you have antivirals on your machines running the docker containers.
And we won't allow to use root just in case you want to disable the av. And we don't care you need it for the docker 😂
This is how I was played when wanted to use docker containers to avoid dealing with OPS.
Some time passes, my team is going to have independent cloud infrastructure.
Doing corporate politics is challenging...5
I'd like to Dockerize our Node.js webapp at work, though I've never worked with any form of containerization software. Has anyone got any good resources for learning it? Should I also be learning Kubernetes?10
I deployed docker on a VPS a few weeks ago as a sort of learning experience since I haven't really worked with containers much before. Today I learned that docker doesn't like firewalls.
Or, to be more specific, it adds rules to iptables that are applied prior to ufw rules, allowing external connections that I really didn't want to allow. If I don't explicitly specify that a port is to be published only to localhost, then it punches a hole through my firewall without telling me.
Which means that all of my containers running behind an nginx reverse proxy that auto-redirects to HTTPS... were also accessible directly via HTTP.
I'm... trying to think of a reason why this kind of default behavior was a good idea, but I'm drawing a blank.
So yesterday i upgraded the packages on my pi, docker being one of them. After that all my containers were down and couldn't be restarted. The error message said the container was already started...
Using the internet i found out other people had that problem too after upgrading to docker 18.02. I ended up removing all containers and searching the bash-history for the docker-run-commands to create new containers.
DOCKER, LISTEN UP. YOU'RE COOL AND SAVED ME A LOT OF TIME. YOU HAVE MINOR ISSUES, BUT THAT'S OK. BUT IF YOU CRASH MY CONTAINERS ONCE MORE I'LL GO CAPTAIN AHAB ON YOU.
Oh, and some more armhf-containers would be nice. WHY CAN I ONLY RUN DOCKER-REGISTRY ON AMD64-ARCHITECTURE?
Docker with nginx-proxy and nginx-proxy-le (Lets Encrypt) is fucking awesome!
I only have to specify environment variables with email and host name when starting new containers with web servers, and the proxy containers will automatically make a proxy to the new container, and generate lets encrypt ssl certificates. I don’t have to lift a fucking finger, it is so ducking genius2
After going through the regular process of talking to HR/Recruitment and passing the casual interview with a team-mate for cultural compatibility, I got the task of grilling a candidate on some technical matters. This being a PHP job, we got to talking about PSRs (PHP Standards Recommendations).
As he seemed to take pride in his knowledge of PSRs, I decided to focus more closely on that.
So we got to a recomendation regarding dependency injection containers. Nothing special, and he seemed to know his stuff. At that point, he made a statement that parts of that recommendation were a bit stupid.
Now, I hate to put people in their place, but his statement did not match what that specific PSR stated. So I gently tried to correct him. The candidate, being on fire thus far, pointed out that I should trust him on this, as he clearly knew his stuff.
Again, I didn't like having to do this, but I also did not like him having a misconception about a topic he was, otherwise, really on top of...
So I asked him to trust *me*, as I was one of the writers who contributed to the standard.
The true test here, of course, wasn't if he knew all the minutia of every standard but how he would react to being corrected.
We, as developers, are wrong all the time. Its how we learn and evolve. So being able to accept that is vital.
Sadly, he did not respond too well and sunk into a bit of a sullen silence. At first I though maybe I'd scared him or that he was afraid of having made a gaff but it soon turned out he genuinly did not like being wrong.
Sadly, I had to advise against hiring him.2
AWS summit, speakers talking about technologies that amazon didn’t build but they provide on its cloud.
All about how it’s awesome to use those technologies on its cloud infrastructure.
Feeling like I’m on some bad advertising summit.
I heard docker, containers like 100s times already.
On one of the slides they claim that 85% of tensorflow workloads are on their cloud.
That’s powerful statement.
Looks like enterprises are all on the way to Oracle 2.0 called AWS.5
!rant - it's a THANK YOU!
Had the problem so far that I could not start some apps in docker containers with GPU support (e. g. Chromium).
After a long search and a lot of help from the community, today's update of xf86-video-ati (1:7.8.0-2 -> 1:7.10.0-1) has finally fixed the bug. Yay!
Thank you very much Arch Linux and all the great maintainers. You're doing an awesome job.2
Containers, specifically Docker and Compose. It's a beautiful model and it solves a ton of real problems, but the list of unforeseen footguns never ends.1
Having a shit of a time trying to figure out why Docker containers are not accessing other containers via domain names as they should technically be going through the jwilder nginx proxy container.
Why can't environment setups ever be simple?
Spent about 5 hours trying to figure out why php-fpm 7.2 was sending a timeout to nginx inside a CentOS 7 container. In the end I couldn't figure it out, downgraded to PHP 7.1 and then everything start working fine. I really hate using CentOS for containers since it also requires privileged mode to be able to start services. Hopefully I can move away from using CentOS for containers soon since the base image is also quite fat.
Some time ago I was looking for materials regarding Ansible automation tool and realized that most of them suggests setting up the lab environment using virtualization (like VirtualBox). In my opinion that is not the best approach – virtual machines consume lots of resources and take some time to start/kill. So I decided to write a guide for setting up Ansible lab environment using Docker containers. Containers require significantly less resources than vm’s and can be bring up and down really quickly. Additional advantage is easy way of automating whole environment using docker-compose. You can find my guide at github: https://github.com/LMtx/... Any feedback is very welcome :)3
I’ve been writing a node.js service that automates kubernete commands so users aren’t having to constantly call for newly generated pod names and then container names and so on so forth.
For instance it would ask you if you want to delete a pod, Exec into a container, fetch logs on a pod. It would list all namespace, pods and containers for those pods for you to pick from by just providing a corresponding number.
With this kind of service assist anybody who is constantly having to restart, delete, fetch logs and delete with Kubernetes?4
So I wanted to do a quick test before going to dinner and now I'm stuck on waiting for this f*cking cloud provider to start my container.
"Provisioning 20 minutes" WHAT THE HELL!? After 20 (TWENTY) minutes my container still hasn't started!?
Is it a joke? Is some sysadmin spying on me and making me wait on purpose? What the f*1
Can somebody explain to me why developers (especially web) have to micromanage every single thing into it's own f*ing component.
Story time: I have an input form with some tabs. I discovered that the UI Library (Devextreme) has a nice little component that handles forms, (including tabs, groups, etc.). So I make a page, configure tabs, inputs and whatnot.
Now, I already knew that my coworkers can't handle html that is bigger than a page. So instead of putting the configs in the frontend, I made nice files where I store those, to keep them nicely clean and seperated.
Me feeling very good, went off to have a nice lunch break.
I come back read the message from my coworker, asking me to make every tab it's own component and form and load them into a separate Tab-Component, instead of using the built in configuration
Like seriously. I have a f*ing library that handles that, why the f*ck do I need to reinvent the wheel here!?
Supposedly it's to make it more maintainable, easier to find bugs, flatten the hierarchy.
Here's a little wake up call you morons: Nesting hundreds of components into each other does *not* help you with that.
It just creates a rabbit-hole of confusing containers that you have to navigate and dissect every time you try to find something.
"Can I fix the bug in the detail Page? Sure I'll tell you tomorrow when I find out which fucking component the bug results from".
Components are there to be *reused*. It's using inheritance for reusing code all over again, but worse.
But maybe I'm just old fashioned, and conservative. Maybe I'm just a really bad software engineer, because nowadays everything seems to result in architectures spreading hundreds of folders, thousands of files with nothing but arbitrary cut-offs with no real benefit, that I don't see the value in.6
Duck! this sloppy whiny winnfsd.
Yay! Let's use state of the art Docker with a VirtualBox VM on Windows10.
Don't get me wrong.
The Docker containers in this VM doing a great job on performance.
But in the very moment a Docker container uses a mounted folder via the windows network filesystem, all hell is breaking loose.
Building a vendor folder using a composer Docker image with 84 Packages takes about 15 seconds when cache has been warmed up.
The same Docker command pointing on a folder mounted to Windows Filesystem with warmed up cache takes about 10 Minutes!@&&@""+&
And what is the duckin' reason for this delay?
Because every transfer of a teeny tiny file has to establish a connection to fat ass Windows OS and has to pass it's glorious "security" layer.
I currently working on a shell script which builds the whole vendor folder on a volume on Docker VM.
After completion, the shell script will compress the folder to one file.
This one file will be transferred over this god damned network filesystem.
Finally the script will unpack the compressed vendor folder in it's destination folder.
What year is it?!??3
Does anyone know why websockets aren't used far and wide for APIs? I mean not like chat applications, but the typical webapp, say an online shop. For me it seems kinda wasteful to fire separate requests with tiny payloads all the time. I currently use moleculer and socketio for a quite big project with multiple websites and backend containers, and so far, i haven't found a disadvantage.
So what have i missed?12
Stop restyling confluence and bitbucket taking more space in an ugly fashion. Don't fix what ain't broken. **#*#*#1
Was working on setting up a ci/CD pipeline. The ci part with automated testing and deployment to a on-premise docker registry worked already, so I thought "hey I could try to actually run one of those fresh containers" so I tried it with the usual docker run command.. "Manifest not found" suddenly appeared, it confused me a bit since I used the same url I used for publishing... So I googled around only to find NOTHING that is even remotely connected to my issue. "Eh let's let the guy that runs that registry fix it" was what I thought and called it a day. The next day I was eager to try it again and checked the urls case by case only to notice that I wrote secret-project-backend-client instead of secret-project-api.. I tried it with the new name and it worked!
Never felt so retarded in my life....
Going to build a blog and use the process as an opportunity to learn a full stack. The question is: LAMP or MEAN or smth else?
I have more (albeit very limited) experience with LAMP so the ecosystem seems a bit more transparent. PHP connected to httpd via PHP-FPM. WordPress with MySQL for the blog posts. Or go flat-file with Statamic. Then shared hosting, FTP, etc. But it's all a bit old school and it seems like the grass is greener and growing faster in the fancy JS garden. And JS for front and back sounds appealing as a learning investment.
But with a JS full stack, I don't even know where to start. I know some basic vanilla JS (about as much as I know PHP). Node, Express, Angular? Never touched 'em. But it seems like if you're gonna dive in and learn some web dev in 2019 these are the things to learn, no? Still more questions: how the hell do you build a blog with JS? Do you need a CMS? Flat-file? And how do you host/deploy that stuff? Droplets, virtual machines, containers? Heroku? Digital Ocean?
Can anybody shed any light? I am not trying to overwhelm myself with a stack too complex for a blog (MEAN), nor shoot my future self in the foot learning fading technologies (LAMP).8
While i do have a homelab, I decided to use my raspi for PiHole and Ubiquiti Cloudkey. Last week I started not being able to access either WebUI, or ssh into it. So today I figured i'd look into it.
Plugged in the hdmi cable, and lo and behold... kernel panic.... Well, guess it time to spin up docker containers on my homelab instead
Why somebody would think that allocate huge amount of objects in the static memory make any sense?? Why??? You need to allocate a bloody database context and all the allocation of your IOC containers and keep increasing!!!
Rails, React, React-Native, Docker, Kubernetes, Openstack, Jenkins, AWS, Microservices, Realm, MongoDB, PostgresQL, GraphQL (list goes on...), and I'm not even done yet.
6 months was spent learning all of the above because I found a Rails-only monolith on Heroku unsettling. My first batch of containers was just deployed and I couldn't be happier. Love my job.3
The problem with moving Docker containers from your decked-out dev machine to a VM on AWS when your boss has told you to keep costs down:
1. Start Micro instance, 1++ gig memory
2. Get Out of Memory error from app after 30 minutes
3. Goto 1
Been trying to learn Docker when I hit a brick wall. How do I use nginx reverse proxy + letsencrypt with multiple containers? I only managed to do it with a single container. Using docker-compose or stuff like that I guess?6
This was an effort to combine Gitlab, Github and Bitbucket with VSCode and git SSH authentication. SSH agent doesn't work, configured, added some code in .bashrc, seems fine. Then there was still ssh-askpass missing.
"ssh_askpass: exec(/usr/bin/ssh-askpass): No such file or directory"
WTF VSCode? Why do I need this crap?
However, installed it. Nevertheless, I'm still asked for my password every time when I synchronize using the GUI. Thank God everything was in docker containers/images. So at least there is no garbage left after every failed attempt.
I don't know how, but I finally made it that at least synchronization using the terminal works without a password.
Took me five hours to do this shit.
Now I just report the bug to Microsoft and then straight to McDonalds. I'm starving.1
So, is it a bad/good idea to run *every* service I want to use on my root server in docker containers?14
Fucking Docker, for no fucking reason (no updates, no changes, etc), I tried booting it up following the morning ritual, and nope, ERR_EMPTY_RESPONSE when connecting to my current project (means I managed to connect but for some reason no data is sent). Nginx container doesn't yell about anything.
Everything around works. Accessing the container works. Even pinging my dev domain works. Why the fuck suddenly fucking Docker just **stops**?!
Restarted Hyper-V, updated laradock, recreated containers, disabled AcrylicDNS. NOPE.
"Fuck you Phlisg, I'm not in the mood today" <lunatic Docker is lunatic>
When your docker project finally runs on Linux containers after two days of debugging because Visual Studio added a \ instead of a / to docker-compose 🙄😅
I am really being tested with my creativity in naming conventions with these two sites rn.
Site 1 is a blog for a place called "The Post", so literally everything is called .post
Site 2 is a development built out of shipping containers, where each container is a different features of the property... just like a container would wrap features in their app🤦♂️3
Does any one know how Containers (?) like Zip or exe files work? I mean when you open it with a texteditor you can't see much?!15
So, I manage my server with docker containers (nginx-proxy and the letsencrypt-companion). I limit access to some subdomains using basic auth, but I want to use client certificates for convenience.
So my questions to the experts:
1) Do you know a good (and convenient) way to manage client certificates ? This should include revoking certs and allowing specific certs only for specific subdomains.
2) Should I use my letsencrypt CA for this or would a self signed CA better suited?
3) Any things I should be aware of?1
Coworker: "Hey, so I discovered this library that automatically brings up and tears down local containers to perform unit tests on data sources"
Me: "Sounds neat"
Coworker: "Yeah, I've been messing with it locally, and it means we don't need to have the data sources installed on our machines or rely on the ones in the testing environment."
Me: "That's good"
Coworker: "Just a shame I had to roll back our testing framework to a previous version and refactor the code in all our other tests as a result."
Me: "Wait what? *looks at documentation* It says they support the newer framework"
Coworker: "Yeah, but I couldn't get it to work. So I'm just gonna make a PR for it, okay?" *Proceeds to make a PR, approve and merge the code before I can comment further on the changes*
Welp, there goes all my motivation to get anything done for the rest of the day.3
One of the weirdest aspects of Docker for me is cross-compiling program installations. One would think that something as complex as a container with several programs that each make unknown decisions based on the environment as part of the installation process can't be cross-compiled.2
I need your input on the following
how important do you think having high core count in CPUs in your daily workflow?
I'm planning on buying a new ryzen 5000 processor, while I am going to game the hell out of it I'm also planning to run wsl2, a ton of chrome tabs, maybe have multiple IDEs for developing random stuff, maybe some virtual machines for some experimentations, some docker containers for some selfhosted software and lastly open demanding games while having everything else open.
Will a 6 core 5600x be enough? or do you think investing in a 5900x will be worth it down the line? (lets say for the next 7 years)
Assume that the GPU will handle the games im going to play and the RAM is going to be 32gb for now12
as a seasoned systems eng myself, i had huge mental block of "i am not a programmer" whining when starting to incorperate agile/infrastructure as code for more seasoned syseng staff.
leadership made devops a role and not a practice so lots of growing pains. was finally able to win them over by asking them to look at how many 'scripts' and 'tools' they wrote to make life easier... and how much simpler and sustainable using puppet/ansible/chef/salt... and checking in all our sacred bin files and only approved 'scripts' would be pushed thru automation tool after post review.
we still are not programmers or developers, but using specific practices and source control took some time but saving us loads of time and gives us ability to actually do engineering
but just have 2 groups of younger guys that grew up wanting to be the bofh/crumudgen get off my systems types that are like not even 30... frustrating as they are the ones that should be more familiar with the shift from strictly ops to some overlap. and the devs that ask for root now that they can launch instances on aws or can launch docker containers and microservice..... ugggg. these 2 groups have never had to rack and stack servers, network gear, storage... just all magic to them because they can start 50 servers with a button click.
try to get past the iam roles, acls, facls, selinux and noshell i have been pushing. bitches.
so i was validating some user data using ajax in a facelets page.nothing was actually being validated when i deploy the application in tomcat..i thought it some sort of a bug on my code and after checking for the greater part of today i just couldnt find any bug.So i decided to deploy on GlassFish and it worked perfectly fine..fuck containers!!!1
Making containers of all your projects and then realizing that the containers are no where near public since they run in thair own bridge adapter 😅1
My website is now deployed on a Digitalocean droplet using Terraform to provision the infrastructure and Ansible to configure the server. It creates users, sets up SSH config and deploys the required containers I want all using an Azure pipeline and an Azure storage account to store the TF state.
Now I need a frontend... ._.2
My answer to their survey -->
What, if anything, do you most _dislike_ about Firebase In-App Messaging?
Come on, have you sit a normal dev, completely new to this push notification thing and ask him to make run a simple app like the flutter firebase_messaging plugin example? For sure you did not oh dear brain dead moron that found his college degree in a Linux magazine 'Ruby special edition'.
Every-f**kin thing about that Firebase is loose end. I read all Medium articles, your utterly soporific documentation that never ends, I am actually running the flutter plugin example firebase_messaging. Nothing works or is referenced correctly: nothing. You really go blind eyes in life... you guys; right? Oh, there is a flimsy workaround in the 100th post under the Github issue number 10 thousand... lets close the crash report. If I did not change 50 meaningless lines in gradle-what-not files to make your brick-of-puke to work, I did not changed a single one.
I dream of you, looking at all those nonsense config files, with cross side eyes and some small but constant sweat, sweat that stinks piss btw, leaving your eyes because you see the end, the absolute total fuckup coming. The day where all that thick stinky shit will become beyond salvation; blurred by infinite uncontrolled and skewed complexity; your creation, your pathetic brain exposed for us all.
For sure I am not the first one to complain... your whole thing, from the first to last quark that constitute it, is irrelevant; a never ending pile of non sense. Someone with all the world contained sabotage determination would not have done lower. Thank you for making me loose hours down deep your shit show. So appreciated.
The setup is: servers, your crap-as-a-service and some mobile devices. For Christ sake, sending 100 bytes as a little [ beep beep + 'hello kitty' ] is not fucking rocket science. Yet you fuckin push it to be a grinding task ... for eternity!!!
You know what, you should invent and require another, new, useless key-value called 'Registration API Key Plugin ID Service' that we have to generate and sync on two machines, everyday, using something obscure shit like a 'Gradle terminal'. Maybe also you could deprecate another key, rename another one to make things worst and I propose to choose a new hash function that we have to compile ourselves. A good candidate would be a C buggy source code from some random Github hacker... who has injected some platform dependent SIMD code (he works on PowerPC and have not test on x64); you know, the guy you admire because he is so much more lowlife that you and has all the Pokemon on his desk. Well that guy just finished a really really rapid hash function... over GPU in a server less fashion... we have an API for it. Every new user will gain 3ms for every new key. WOW, Imagine the gain over millions of users!!! Push that in the official pipe fucktard!.. What are you waiting for? Wait, no, change the whole service name and infrastructure. Move everything to CLSG (cloud lambda service ... by Google); that is it, brilliant!
And Oh, yeah, to secure the whole void, bury the doc for the new hash under 3000 words, lost between v2, v1 and some other deprecated doc that also have 3000 and are still first result on Google. Finally I think about it, let go the doc, fuck it... a tutorial, for 'weak ass' right.
One last thing, rewrite all your tech in the latest new in house language, split everything in 'femto services' => ( one assembly operation by OS process ) and finally cramp all those in containers... Agile, for sure it has to be Agile. Users will really appreciate the improvements of your mandatory service.
LXC/LXD containers are awesome for Windows VirtualBox users. It would allow creating a single Linux VirtualBox VM and then create multiple LXC containers (full-blown Linux machines) utilizing all the full resources of host Linux VirtualBox VM, keeping it always a clean image. Super efficient utilization of RAM and storage. No need to create multiple VMs for different Linux OSs.2
!rant but sometime you need to share some positive vibes.
Found out I could get $50 credit for digital ocean from github because I am a student.
So now I can learn a lot for free, and if I mess something up I can just create a new machine.
So now I am first learning how to work with docker and the communication between containers.
Good to see people want to encourage devs :)2
Windows machine i5 7th gen with 12Gigs Ram and 4Gigs graphics (2 dedicated + 2 shared) and 1TB HDD+ 256 Gb SSD vs low budget MacMiNi for heavy dev? usage- JDK, Android SDK, node, containers etc.9
So I wanted to use aws-localstack and it's ok I guess. Except who built these docker containers? Ever heard of SRP? A single-monolithic container with all services inside, glued together with python and makefiles, using non-standard ports, and on top of that it's hard to configure.
Who wrote this garbage? Atlassian. Oh. That makes perfect sense.
Fuck this, I'm gonna write my own localstack. What do you guys think of this?
Midnight fucks to docker again for flushing my firewall rules and for crashing containers when I add them back just so that my services don't get exposed to all retarded script kiddies probing ports and services. Nice going.4
Tired of installing a shit ton of cli tools, go, npm, whatever locally? Give it a shot and run your commands in project specific docker containers. Also works in Ci.
Im new to golang so I'm also happy about feedback :)8
What is the best laptop for development?
Especially if you work a lot with virtualization, containers, the cloud, and a lot of systems programming. And you like to simulate a lot of distributed environments on locally?3
Every time (by which I mean 2 times so far) I update docker something stops working in one of my containers. Is this common or just my incompetence?2
My company don't want to buy me a new computer: Im stuck with 2012 Dell. No kidding. And I am supposed to run 10+ docker containers all the time, plus java IDE.
Of course, I bought a new computer - by myself. Fuck my company. I don't have time for stupidity.3
I have actually done nothing today. Had a thought that I might puppetize our servers, but I can't install any vm software or even puppet on my pc. I then had an idea to have our CI tool package our apps into docker containers so they could be distributed to workstations and servers alike. But I can't install docker. Sigh.1
I read learncpp.com and in parallel i watched a lot of the tutorials by MakingGamesWithBen, which helped me a lot. This basically thought me C++ in less than half a year and since then i am gathering experiences and i never have problems with the language. All i have to do is look up more specific stuff like special containers or functions
So, yesterday my colleague decided to upgrade his OS from Ubuntu 14.04 to 16.04.
He was working on some projects within containers. Stopped the containers and started upgrade as it was EOD.
Guess what, the update failed and it messed up his system. Now the system boots but crashes after few min. Docker won't start.4
For me that would be Proxmox. I know, people like it - but for no apparent reason it decided to nuke half my ZFS datasets in a pool, with no logic behind it whatsoever. All disks were tested, all came out good. Within the same pool there were datasets that were lost and some that remained.
I really don't get it. Looking at Proxmox' source code, it's more or less the command line tools and then there's the web interface (e.g. https://github.com/proxmox/...). Oh and they have the audacity to use their own file extension. Why not I guess?
Anyway, half my data was gone. I couldn't tell how or why or what the fuck even happened there. But Proxmox runs Debian underneath and I've been rather pissed about Proxmox' idea of "don't touch the host system aaa" for a while at that point. So I figured, fuck it I'll just take pure Debian then and write my own slightly better garbage on top of that. And as such the distribution project was born. I've been working on it for a little over a year now. And I've never had such issues again.
I somewhat get the idea of "don't touch the host" now, but still not quite. Yes, the more you do in the containers, the better. And the less you do on the host in terms of reconfiguration, the longer it will stay alive for. That goes for any system - more reconfiguration means usually means less stability and harder to replace. But sometimes you just have to work from the host. Like say migrating a container between hosts, which my code can do. You can't do that from a container, at all. There are good reasons to work with the host. Proxmox isn't telling that. Do they expect their users to be idiots? Only enterprise sysadmins amirite?
So yeah, that project - while I do take inspiration from it in mine - I don't like it. It's enterprise, it has the ZFS and the Ceph and the LXC and the VM's - woohoo! Not like anyone could implement that on a base Debian system. But they have the configuration database (pmxcfs), the distributed configuration database of a couple MB large and capped there, woah!
Ok sure it isn't Microsoft or IBM or Oracle or whatever, and those are definitely worse. But those are usually vendor lock-ins.. I avoid those on that premise alone :)3
I’m trying to update a job posting so that it’s not complete BS and deters juniors from applying... but honestly this is so tough... no wonder these posting get so much bs in them...
Maybe devRant community can help be tackle this conundrum.
I am looking for a junior ml engineer. Basically somebody I can offload a bunch of easy menial tasks like “helping data scientists debug their docker containers”, “integrating with 3rd party REST APIs some of our models for governance”, “extend/debug our ci”, “write some preprocessing functions for raw data”. I’m not expecting the person to know any of the tech we are using, but they should at least be competent enough to google what “docker is” or how GitHub actions work. I’ll be reviewing their work anyhow. Also the person should be able to speak to data scientists on topics relating to accuracy metrics and mode inputs/outputs (not so much the deep-end of how the models work).
In my opinion i need either a “mathy person who loves to code” (like me) or a “techy person who’s interested in data science”.
What do you think is a reasonable request for credentials/experience?5
Email-server: Installation directly on the server vs docker-containers. Does anyone here have positive/negative experiences with any of them?10
So this week should be interesting. I am working on a (potentially) very large project for my current client and need to build a service that somewhat replicates the functionality of heroku (in that it needs to be able to load an app built in one of several languages, and spin it up in a docker container).
Unlike Heroku, however, each application also needs to be able to have a list of public and private (internal only) API routes listed and be able to dynamically route requests to the correct routes on in those containers. (Sorry if this is confusing)
Does this sound challenging and amazing? Absolutely! Do I think I may be in over my head? Yes, yes I do.
Has anyone ever built or worked with something similar?1
Docker pisses you off? Everything worked fine and suddenly doesn't?
$ docker system purge
I like docker. I think the technology is cool and I have a few ideas on how I can use it. But I sometimes think I'm too stupid to use it.
I'm currently trying to convert parts of my dev environment to use containers. But following tutorials on it confuses me and I get lost with it really quickly and get discouraged for a few days until I give it another try and fail again
Like fuck I already know I'm mediocre but this makes me feel like i should go into a manual labor field instead of software development6
I'm running unintuitive server on a Mac mini for home Dev type stuff and it's running warm but quick. Has two virtual containers innit as well - admittedly I'm not banging it hard but seems fine.
Would be interested in other views is Linux server as on Linode Ubuntu seems to be standard1
Is there any kind of protocol/method where I can use something like docker containers in order to "host" compilers like gcc and use that with vscode to compile and assemble source code?
No I'm not talking about volumes (it's a bit tedious if I want to use it to manage numerous projects)3
Is it just me or others too feel that docker completely nailed it with their naming convention.
Like docker, containers etc are so aptly named by their functionality.1
Has anyone had issues with the Docker Go Client (moby) and using containers with pointers? It would seem that the client swaps pointers or something, so I'd have to use the value instead. Might be very unclear, but if someone's had similar problems, I think you'd understand 😅1
Career fair was today at my college. I don’t know why but I makes me kinda happy that some things I’m learning on my own were mentioned that aren’t taught in class
Many people mentioned micro services, some talked about containers like docker, and I had a ten minute conversation just about react js. All things I’m reading and toying with on my own time since they’re not in the curriculum, but I’ve heard them mentioned often enough in articles to know I should know about them
Why the heck does everyone thnk it's a good idea to run docker Containers inside a vm?
Or Containers in general...
This is unneeded redundancy and it kills the performance aspect, which makes containers favorable in comparison to vms...3
I have done some experiments on my server in the past. It's a great way to learn new things. However, I am bound to make some mistakes and over time the sever becomes messier and messier.
A week ago I installed UNRAID on my machine and I love it! I can now have my critical infrastructure live and working in docker containers and vms.
Then if I want to do an experiment I spin up a VM in a couple of minutes to do my thing and remove it when I am done. No traces left!
I just did *skadoosh* while executing command to run multiple docker containers with docker-compose and pressing the Enter (Return) key with my pinky finger.
Feels like real-life Kung Fu Panda trying to do a Wuxi Finger Hold.
I bet you will never forget this when you're running a command on laptop and pressing enter with pinky finger.3
Can you start services manually when using docker compose?
As in having container A, B and C with an executable A, B and C (originating from the same source btw) where I can start/stop A, B, C independently?
All examples seem to use it as a whole thing where all containers must be up for the service to work.3
Is docker even suitable for anything that isn't deployment?
So much time, so much effort, so much trial and error, and I still feel like I don't know what Docker is for.
I had a development VirtualBox machine, which I used just to compile my code and test my application. So I said "why don't I just use Docker? It would be way simpler". Also because that fucking Virtualbox image was like 10GB, and it was slow af.
The VirtualBox machine wasn't created by me, but it was just given to me by a previous developer, so I just had to imagine what I needed and pick up the pieces. In few hours I was ready with my Dockerfile.
So I tried it, and....... obviously it didn't work. I entered inside my container and I tried to manually execute commands in order to see where it breaks, and I tried to fix each of them. They were just the usual Linux dependencies problems, incompatibility among libraries, and so on.
Putting everything in order, I started over again with a virgin Ubuntu image, and I tried to fix every single error that appeared, I typed something like 1 hundred commands just to have my development machine up and running.
Now I have a running container that works, I don't know how to reproduce it with a Dockerfile, and I don't know what I'm supposed to do with it, because I'm afraid that any wrong command could destroy the container and lose all the job I did. I can't even bind folders because start/exec doesn't support bindings, so I've to copy files.
Furthermore, the documentation about start/exec is very limited, and every question on StackOverflow just talks about deployment. So am I wrong? Did I use containers for something that wasn't their main purpose? What am I supposed to do now? I'm lost, I feel so much stupid.
Just tell me what to do or call a psychologist8
The feeling you get when you have to refactor a 'refactoring effort of a common library for a better unit test coverage', because someone forgot to factor in apps using dependency injection.
DI containers have feelings too!
Making distributed scheduler that queue and run tasks on containers or other executors in future and also pulls new tasks from defined git repositories.
Tasks are added based on simple yaml configuration.
Need that for my side projects that gather data from multiple sources from time to time.
k8s looks to heavy for that and airflow can’t be configured like I want it to be so I started writing my own on Monday.
Nearly finished poc version.2
Before my area of focus was the frontend and then my boss thought my skills were good enough to handle servers, now I'm here and working on the staging and production containers... Mother docker awesome...6
Jesus Christ, Docker Desktop for Mac is fucking garbage.
Did you start your containers with an alternate compose file? Well, fuck you, we are goddamn idiots who don't know what the fuck we are doing and our piece of shit GUI just doesn't know how the fuck to deal with that.1
I somehow managed to create 2 different IDs and realize that I'd need to store previous value for one of them - in a tabbed view component. Although to be fair the component allows reordering tabs and moving them across containers with drag&drop.
Hi everyone, I’m trying to wrap my head around dev ops but struggling with the whole continuous integration workflow. From my understanding it goes something like the following:
1. pushing a change to some repository (git)
2. Some tool (Jenkins) tries to build it and if it’s succeed, creates a image /container (docker).
These containers are hosted on some cloud service (aws)
Some workflow, walkthrough, or examples would be very appreciate.7
Why am i just now looking into linux containers?! Would have made life so much easier and kept my server less messy and shit!
Anyone can tell me the pros and cons of docker, rkt from coreOS, and LXC?
Spent ages building docker containers for all our projects. All worked great and I released it. So someone added a new dependency that required an Ubuntu package to be installed but forgot to tell me. My colleague said "docker isn't very stable is it" when it didn't run due to said dependency. It's a Linux container ffs, see if you blame Virtualbox under the same circumstances you idiot
Oh there are lots of good tools to handle different screen sizes like containers, anchors, scaling and all that good stuff huh ?
Sets it to stretch mode instead
Who could explain what a container is? Like a Docker container, but not necessarily Docker. Google doesn't provide simple yet meaningful answer (or there is something wrong with me today). Any ideas?8
Is relying on probability bad practice? I have a container that needs to know about all other instances of the same container and assign a unique ID to itself on initialization. I thought that if I selected a big random number as its id on startup, the chances of it having the same id as another instance is very low.
To prevent two instances from having the same id I could check for all running instances on the network, but what if two instances start at the same time? They won't find each other since none of them will be fully initialized when their id checks run.
Probability says I'll be fine, Murphy's law says I won't. What do?5
Any recommendations for decent 13 inch laptop under 1400$ USD
- Atleast 8 gigs of RAM and option to upgrade
- resolution >= full HD
- storage >= 256 GB SSD
- battery >= 6 hrs heavy chrome usage
- Linux compatibility
- Cpu- nothing heavy - browser and docker containers, no building or compiling17
I was setting up a CI build machine. Builds were supposed to be ran in disposable containers, but I needed a way to trigger a task on the host from inside the container. I didn't want to give containers shell access to host - kinda misses the point.
Solution: a server running on the host and listening for predefined commands on a named pipe. The pipe was bound into containers which would simply echo commands into it. Very simple and effective.
The hacky part? The server was an 8-line bash script.1
Note to self:
Variadic C++ templates combined with obfuscated combinations of stl containers put you on the right path to be a "compiler message decryption" archmage.
Especially when you use MSCV...
I've got a somewhat special issue with my setup.
I am running an instance of `lucaslorentz/ caddy-docker-proxy` as proxy that handles certificates and request and proxies them to docker containers that run `abiosoft/caddy:php` to host Laravel based applications. The problem is, that the `abiosoft/caddy` containers do not know it's assigned hostname and thus Laravel's `asset`, `secure_asset` and `url` respectively `secure_url` don't work as they use the internal hostname which would be an IP address and thus requests go to 192.168.240.x instead of example.com.
I am not yet entirely sure where I should tackle this problem and am grateful for every hint.
I am currently also evaluating traefik instead of Caddy-docker-Proxy and Caddys v2 official container instead of abiosoft's Caddy v1 container but I guess, that this wouldn't solve the issue as the container still wouldn't know that it's given Domainname is example.com4
Does somebody have any experience with LXC/LXD containers on servers? I basicaly want all the services separated, but still have an easy way to manage the networking/routing for all the services and containers.
Any reference, guide or tool that helped you in master this subject?
Thank you in advance!4
So at work, we need to order an ECS to containerize our apps.
They're asking how much RAM and mili-cores and I have no idea about what that means.
Seems you can request a factory cou but what does that mean. If I have a 3.2ghz 1 core CPU for example. And I ask for 100 mili-cores... What does that mean?3
recently I've been thinking about a good way to deploy mass amounts of containers on kubernetes, but before I just go with docker, I want to learn about some alternatives
what are some other container softwares you guys use?2
Do we still need virtualenv when we have containers? Cuz one of my friends thinks it’s a good idea to use virtualenv everywhere, even in production, even in an LXC?
Is it being paranoid or really a solution to some problem that would otherwise rise? I’m just curious.1
Finally got around to migrating my two servers after Scaleway and Aruba raised their prices... man these migrations are a pain in the ass.
And I *still* have not finished migrating the docker containers, because that piece of trash still eludes me.
Maybe it's my depression speaking, but this shit is exhausting2
I was finally able to set up a server with my coworker's help in our work office. After struggling trying to get CoreOS to work, I quit and just went with Ubuntu Server 16.04. It's a SFF PC that we had lying around. The purpose is to run applications in Docker containers for QA, demo, and performance testing. I can say it was truly a productive day...
I got my first developer job three years ago. I’ve always had a great eye for detail, and getting things done while following best practices. I learned that a few years ago from typography, which I think is a fascinating subject, which has a lot of shared ideas with software development.
In my first job, I immediately took a lot more responsibility than what I was assigned to. This job was as a React Developer, but I quickly got into backend development and set up kubernetes clusters, CI/CD.
Looking back, this was to me quite an achievement, considering I had never done anything even remotely close to it.
I did however, work my ass off. 18 hours work days without telling my boss, so only getting paid for 8. Plus I worked weekends.
I did love it. After a while, I got promotes to Senior Developer, and got responsibility for everything technical. I tried asking for help, but everybody else was either a student, or working purely front-end or app-development. Meanwhile, I was Devops, API-design, backend, Ci/CD, handling remote installations (all our customers are Airgapped), customer support, front-end and occasionally app-development when the app-developers could not handle their shit. Basically, I was the goto-guy for every problem, every feature, every fix. I don’t say this to brag.
I recently quit my job, started working as a consultant, because I almost doubled my pay. However the new job is boring as shit. I’m now an overpaid React Developer. And I really hate React. Not because it is shit, but simply because it is boring.
I’m thinking of going back to my old job. It was a lot of work, but it was really interesting. However, after I quit, they have changed their whole stack. No more Golang, Containers, Kubernetes, webRTC and other fun new technologies. Now, it is just plain, PHP without any dependecies. It is both boring, and idiotic. So I’m thinking of just quitting. Either doing some personal projects like game-development. I dont know.
Why in the fuck does everyone expose specific ports in Dockerfiles?
If I wanted to expose the port, I would fucking expose it.
Currently can't run my home infra platform because I'm running two separate instances of Maria DB on the same private internal network. These are two databases for two separate applications.
Why don't I run them on one? Because they're two separate fucking applications.
Why the fuck can I not do this when I used to be able to do it a week ago.
Stop exposing your fucking ports in your fucking Dockerfiles.
This shit is getting so bad, I'm just about to throw my towel in on all fucking containers and just install everything in multiple VM environments.
I am God damn appalled that after 8 years of using docker, core concepts like a port exposure is being leveraged as a way to somehow circumvent poor security practices.
You want a secure container environment? Expose your own goddamn ports.
Fuck you Maria DB, and fuck you docker.2
Setting up dockerfile with ENV(from Visual Studio) was such a stressful endeavour from the point of view of someone that doesn't work daily with containers that I'm wondering how you master folks of net core and docket live and breath.
The gotcha was to put the ENVs at the first FROM from which the running environment was going to execute the WebAPI and not later on where there is the ENTRYPOINT point
Every once in a while, when I have forgotten about last time debugging and pruning docker leftovers, I have to stop and kill images. Once I'm sure everything has been stopped, trying to remove outdated volumes, getting "volume is in use".
Never seems to make any sense to me in that moment. How can it be in use, when no process should be running that uses it?
Then I remember, apart from containers and volumes, there are images ...
If someone can shed some light on this behavior, would be appreciated:
I am running a couple of docker containers with lighttpd on my server (lighttpd is also installed on the host server for reverse proxy). Now whenever I kill lighttpd on my host server it also kills ALL the running lighttpd instances in my docker containers. Isn't docker supposed to be, idk, CONTAINERIZED?2
Bit of a stupid oopsie I had today that someone might appreciate.
We’re working on a microservice project in Spring Boot, running in a docker swarm. Past few days I get a Spring Cloud config server going in separate stack, create an overlay network, and get CI deployments to use the right profiles etc. It’s looking great, and the first component is working spectacularly.
Now just to do the other 6. Move config files to the Git repo, tweak CI, all the other faffing and hoohas; and deploy. Health checks keep failing, the containers are murdering themselves and resurrecting ad infinitum. They’re doing this so quickly that by the time I get the container ID to exec in and curl health, it’s no longer running. Cue frustration, increased caffeine and nicotine consumption; my sanity is slipping.
No errors in the logs, because from experience the Cloud Config errors ar at debug level. Whhhyyyy?? Some time later (way longer than it should have been) I realize I had never actually included the Spring Cloud Config starter. Boot 101, get your starter!
Since config client is just additional setup in properties.yml, there’s no issue of the dep isn’t there, it just doesn’t try to get the config.
The containers are still unhealthy, I can hear them screaming. But now at least it’s about something else...
Okay so i am trying my hands in web dev, majorly the backend stuff and my aim is to launch a simple web app in a production grade environment, to get a basic knowledge of various tools.
So i had the following project in mind: "save my secret string" . this is going to be a minimal site with 2 blocks.
in the first block in which user can add an id_string and data_string, and press enter. on pressing enter, the data will get saved in backend. if the id is already used, it will over write it with new data.
in the second block there's going to be a search button and an output area where user can enter the id string and get output( no matter who added the output, if key is correct, they can access it) if its available in the database. Kind of like a pastebin, but without any links
is it a simple project to learn the back end basics?I plan to add more features to it and later convert it to full scale blog site.
So can you help with a few pointers on what should i be looking upon to get it running like a full production app ? I know of a few buzz words that production apps use in their environment: docker multi containers, vanilla html/css/js , some form of encryption lib, aws instances and vpc / multi user ec2, mysql, postman and some kind of js framework for sql queries( i guess that's postgres)
I don't really know much about any of them, so can you help me with a flow on what should i learn to make this app possible?