Do all the things like ++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatarSign Up
Get a devDuck
Rubber duck debugging has never been so cute! Get your favorite coding language devDuckBuy Now
Search - "containers"
So i heard that docker is now a thing and decided to put my servers into containers as well (see image)...but somehow I still don't understand how is this supposed to simplify deployment, vertical scaling and so on...? :D6
enabling firewall on a vps to secure my docker containers and forgetting to add openssh to allowed list --> ssh blocked 😃🔫24
That awkward moment when I was able to run three docker containers on a 512MB server:
1. DotNet core web service
BUT I cannot run:
1. NodeJs web service
2. MongoDB container
Spent two hours configuring the damn server to get hit by this T_T14
Biggest rivalries right now ..
Windows vs Linux,
Chrome vs Firefox,
VMs vs Containers,
Vim vs Nano,
SQL vs NoSQL,
React.js vs Vue.js,
Tensorflow vs Pytorch 😎😎19
Soooo my little encryption tool makes progress. <3
After a short break from development, we had our first successful loaded container yesterday!
- Protocoll is working
- We can create containers and store/copress files in it
- we're awesome
- I love it
- you are awesome, too!
(Loaded containers will be inaccessible for movement to different directories while our tool is open)49
So this client wanted a demo on Dockers. So I gave the demo with some microservices running on different containers. Later the clients come back and say, "Docker is good. But please fit all the microservices in one container." I say but that defeats the purpose of microservices. But no, the client say. I tried explaining but no is no. Shit!! Fine! Have it your way!!5
This rant is particularly directed at web designers, front-end developers. If you match that, please do take a few minutes to read it, and read it once again.
Web 2.0. It's something that I hate. Particularly because the directive amongst webdesigners seems to be "client has plenty of resources anyway, and if they don't, they'll buy more anyway". I'd like to debunk that with an analogy that I've been thinking about for a while.
I've got one server in my home, with 8GB of RAM, 4 cores and ~4TB of storage. On it I'm running Proxmox, which is currently using about 4GB of RAM for about a dozen VM's and LXC containers. The VM's take the most RAM by far, while the LXC's are just glorified chroots (which nonetheless I find very intriguing due to their ability to run unprivileged). Average LXC takes just 60MB RAM, the amount for an init, the shell and the service(s) running in this LXC. Just like a chroot, but better.
On that host I expect to be able to run about 20-30 guests at this rate. On 4 cores and 8GB RAM. More extensive migration to LXC will improve this number over time. However, I'd like to go further. Once I've been able to build a Linux which was just a kernel and busybox, backed by the musl C library. The thing consumed only 13MB of RAM, which was a VM with its whole 13MB of RAM consumption being dedicated entirely to the kernel. I could probably optimize it further with modularization, but at the time I didn't due to its experimental nature. On a chroot, the kernel of the host is used, meaning that said setup in a chroot would border near the kB's of RAM consumption. The busybox shell would be its most important RAM consumer, which is negligible.
I don't want to settle with 20-30 VM's. I want to settle with hundreds or even thousands of LXC's on 8GB of RAM, as I've seen first-hand with my own builds that it's possible. That's something that's very important in webdesign. Browsers aren't all that different. More often than not, your website will share its resources with about 50-100 other tabs, because users forget to close their old tabs, are power users, looking things up on Stack Overflow, or whatever. Therefore that 8GB of RAM now reduces itself to about 80MB only. And then you've got modern web browsers which allocate their own process for each tab (at a certain amount, it seems to be limited at about 20-30 processes, but still).. and all of its memory required to render yours is duplicated into your designated 80MB. Let's say that 10MB is available for the website at most. This is a very liberal amount for a webserver to deal with per request, so let's stick with that, although in reality it'd probably be less.
10MB, the available RAM for the website you're trying to show. Of course, the total RAM of the user is comparatively huge, but your own chunk is much smaller than that. Optimization is key. Does your website really need that amount? In third-world countries where the internet bandwidth is still in the order of kB/s, 10MB is *very* liberal. Back in 2014 when I got into technology and webdesign, there was this rule of thumb that 7 seconds is usually when visitors click away. That'd translate into.. let's say, 10kB/s for third-world countries? 7 seconds makes that 70kB of available network bandwidth.
Web 2.0, taking 30+ seconds to load a web page, even on a broadband connection? Totally ridiculous. Make your website as fast as it can be, after all you're playing along with 50-100 other tabs. The faster, the better. The more lightweight, the better. If at all possible, please pursue this goal and make the Web a better place. Efficiency matters.8
DISCLAIMER: UNPOPULAR OPINION
I'm tired of the Linux community, they effectively discourage me of taking part in any discussion online
I'm currently making Windows-only soft, some game stuff, some legacy DirectX stuff you got it.
Everytime I go online, this shitty pattern happens, when I stumble upon a problem in project I don't know how to fix and I ask for help
These are responses
- HA, HA, WINDOWS BAD, HA, HA, GET REAL SYSTEM
- In Linux, we can do X too. I mean it has 4x less functionality and way shittier UX and is even harder to implement but it can probably work on too Linux, so it's better, yes, just move to Linux
- btw you didn't like Linux before? Try this distro man, it's better <links random distro>
Is there anything valuable in the Linux community? I feel like these people don't like Linux anyway, they just hate Windows. Every opinion, tip is always opinion based. Anyone who works on internals knows how much better and how well thought is Windows kernel compared to Linux kernel. Also, if someone unironically uses Linux distro on desktop PC then he's a masochist because desktop Linux is dieing. So many distros ceased work only this year.
Is it a good tool for servers and docker containers? I don't have my head stuck up my ass to admit that yes, it's much better than Windows here.
This community got me stressed right now, I fear that when I go to bathroom or open my microwave there's gonna be a Linux distro recommendation there
Now seriously, WHAT THE FUCK???
Every single time I have to work with people from a particular country [you have one guess. Yepp, that's the one], I see A-FUCKING-LOOOOOOT of manual work?!?
"can you reboot the server?"
-"sure, let me help you, sir" <20 minutes later> "done"
"can you unlock my account?"
-"yes, just a moment sir" <20 minutes later> "please check now"
"can you restart this environment w/ 200 instances?"
-"yes sir, let me check" <6 hours later> "please check now"
"you've missed 18 containers"
-"oh okay sir, will restart them now" <2hours later> "please check now"
[I am already OoO]
why is it that every time I have to work with you guys I am the one who is automating shit. How come you never think of/do any automata? You are fucking technitians, you should know how. WHY DO YOU ENJOY CLICKING ALL-DAY-LONG????
I'm serious. Why??? I'm struggling to understand...25
Are you using socat?
Any interesting use case you would like to share?
I am using it to create fake / proxy docker containers for network testing.7
I'm just amazed what 512MB of RAM can do :O
That's htop from my VPS I feel sorry for the CPU though.
It is running three docker containers:
1. Dotnet Core
Our client decided to save some $$. At the end of each business day teams downscale their environments before leaving and the next day scale them up in the morning to start working.
The idea is not bad, but they are a bit too ignorant to the fact that some environments are exceeding AWS APIs limits already (huge, HUGE accounts, huge environments, each env easily exceeding /26 netmask, not even taking containers into account). Sooo... scaling up might take a while. Take today for example:
- come in to the office at 7
- start scaling up
- have lunch
- ~15:00 scaleup has finished
- one component is not working, escalating respective folks to fix them
- ~17:00 env is ready for work
- 17:01 initiate scaledown process and go home
Sounds like a hell of a productive day!!!
Go home grafana, you're drunk....
Maybe I shouldn't run 50 containers on a system with 2 cores and 6 GB RAM.4
Okay guys, this is it!
Today was my final day at my current employer. I am on vacation next week, and will return to my previous employer on January the 2nd.
So I am going back to full time C/C++ coding on Linux. My machines will, once again, all have Gentoo Linux on them, while the servers run Debian. (Or Devuan if I can help it.)
So what have I learned in my 15 months stint as a C++ Qt5 developer on Windows 10 using Visual Studio 2017?
1. VS2017 is the best ever.
Although I am a Linux guy, I have owned all Visual C++/Studio versions since Visual C++ 6 (1999) - if only to use for cross-platform projects in a Windows VM.
2. I love Qt5, even on Windows!
And QtDesigner is a far better tool than I thought. On Linux I rarely had to design GUIs, so I was happily surprised.
3. GUI apps are always inferior to CLI.
Whenever a collegue of mine and me had worked on the same parts in the same libraries, and hit the inevitable merge conflict resolving session, we played a game: Who would push first? Him, with TortoiseGit and BeyondCompare? Or me, with MinTTY and kdiff3?
Surprise! I always won! 😁
4. Only shortly into Application Development for Windows with Visual Studio, I started to miss the fun it is to code on Linux for Linux.
No matter how much I like VS2017, I really miss Code::Blocks!
5. Big software suites (2,792 files) are interesting, but I prefer libraries and frameworks to work on.
For future reference, I'll answer a possible question I may have in the future about Windows 10: What did I use to mod/pimp it?
1. 7+ Taskbar Tweaker
3. Classic Start (Now: Open-Shell-Menu)
Enhanced text editor I like a lot more than notepad++. Aaaand it has a "vim-mode". 👍
Three way diff viewer, that can resolve most merge conflicts on its own. Its keyboard shortcuts (ctrl-1|2|3 ; ctrl-PgDn) let you fly through your files.
8. Link Shell Extensions
Support hard links, symbolic links, junctions and much more right from the explorer via right-click-menu.
Neither as beautiful as Conky, nor as easy to configure or flexible. But it does its job.
Of course this wasn't everything. I also pimped Visual Studio quite heavily. Sam question from my future self: What did I do?
1 AStyle Extension
2 Better Comments
Simple patche to make different comment styles look different. Like obsolete ones being showed striked through, or important ones in bold red and such stuff.
4 Atomineer Pro Documentation
Alright, it is commercial. But there is not another tool that can keep doxygen style comments updated. Without this, you have to do it by hand.
5 Highlight all occurrences of selected word++
Select a word, and all similar get highlighted. VS could do this on its own, but is restricted to keywords.
6 Hot Commands for Visual Studio
This ingenious invention colorizes brackets (aka "Rainbow brackets") and makes their inner space visible on demand. Very useful if you have to deal with complex flows.
Come on! 2018 and Visual Studio still outputs monochromatically?
That's it, folks.
No matter how much fun it will be to do full time Linux C/C++ coding, and reverse engineering of WORM file systems and proprietary containers and databases, the thing I am most looking forward to is quite mundane: I can do what the fuck I want!
Being stuck in a project? No problem, any of my own projects is just a 'git clone' away. (Or fetch/pull more likely... 😜)
Here I am leaving a place where gitlab.com, github.com and sourceforge.net are blocked.
But I will also miss my collegues here. I know it.
Well, part of the game I guess?7
So we started looking into docker. As always I needed to do the research and I was fine with it.
We have 4 projects that are sold into one suite so logically I follow the microservices build structure.
3 months later after everything has been set up, we get called into a meeting. The whole suite should be a monolith as microservices doesn't make sense to the people planning everything.
Ok pulled my current plans out abd made everything a monolith. Just note I also get pulled away to other Business Units to do work for them.
Get pulled into another meeting 2 months later. Why isn't the docker containers in microservices!? It is stupid running as a monolith and we should've done our jobs better etc...
After the meeting my manager and I just sighed and walked to the office. So basically 5 months doing the the exact same thing we did in 3 weeks.
Now they want to develop other services and want to strip every method into a microservice and bundle it together.
Life of a DevOps engineer right!1
Sysadmin: Apps on containers and kube is mandatory from now on, scaling is mandatory!
Devs: The systems weren’t designed for containers, we haven’t prepared shirt for scaling!
Sysadmin: Hold my beer!
FUCK capitalist greed! I have befallen to their tricks once again. The daily dosage on my gummy vitamins was three a day but the total gummies in the container wasn't divisible by three so I had to buy three containers and eat one from each per day!21
"WTF? These records should have been inserted into the table!"
...Hours of checking code, trying to figure out how this is possible, can't find a way to have this scenario happen...
...Add additional debug and troubleshooting code, add more verbose logging, redeploy to all the containers, reset all the tables, many apologies to the boss for the delay....
...Co-worker comes in: "oh, hey, sorry, accidently deleted some stuff from the database last night before i left."1
I like to come up with superhero names for my docker containers. That makes me feel like the Uber-villain when I run "docker kill elasticman"4
Seriously, Ubuntu can go burn in hell far as I care.
I've spent the better part of my morning attempting to set it up to run with the correct Nvidia drivers, Cuda and various other packages I need for my ML-Thesis.
After countless random freezes, updates,. Downgrades and god-knows-what, I'm going back to Windows 10 (yes, you read that right). It's not perfect but at least I don't have to battle with my laptop to get it running. The only thing which REALLY bothers me about it is the lack of GPU pass-through, meaning running local docker containers rely solely on the CPU. In itself not a huge issue if only I didn't NEED THE GOD DAMN GPU FOR THE TRAINING21
Between containers, cluster managers and virtual machines we've lost track of where our code even is.4
Actually I'm pleasantly surprised about Windows' stability nowadays. It's capable for running for up to a week with no stability issues, whereas systemd on the other hand.. let's just say that my Arch containers could do better right now.
Data mining aside, damn man.. Microsoft is improving for once! Is this the so-many'th unusable/somewhat stable switch? I mean, it's not like we haven't seen that happen yet! Windows 98, shit! Windows 2000, kinda alright! Windows Me, shit! Windows XP, kinda alright! Windows Vista, oh don't even get me started on that pile of garbage! Windows 7, again kinda okay! Windows 8, WHERE THE FUCK DID THAT START MENU GO YOU MOTHERFUCKERS?!!! Windows 10, well at least that Start menu got fixed. Then it got into some severe QA issues, which now seem to have gotten somewhat fixed again.
I'm starting to see a pattern here! 🤔13
Hololens development forced me into Visual Studio after spending years doing Unity development with MonoDevelop in MacOS.
Why haven't anyone told me to switch sooner! Thanks to Visual Studio + ReSharper, my brain farts turn into a coherent code almost automatically.
I hate that I need MacOS for the iOS development and Win 10 for Hololens. Running Win 10 on Parallels kinda works, but it is a compromise. Developing without headphones/earplugs is out of the question if you don't want to go deaf.
I wan't all the tools for a single OS so I don't have to maintain multiple computers and even more importantly travel with multiple laptops. Just love the security check question "Do you have any electronics with you? Please put it into the container." - "Could I get a couple more containers, please..."11
+ Death of Wix and WordPress (hopefully)
+ Abandonment of Java
- Software engineers trying to automate everything
+ Cloud blockchain AI containers that mine bitcoin with machine learning technology in high demand2
So this one made me create an account on here...
At work, there's a feature of our application that allows the user to design something (keeping it vague on purpose) and to request a 3D render of their creation.
Working with dynamically positioned objects, textures and such, errors are bound to happen. That's why we implemented a bug report feature.
We have a small team tasked with monitoring the bug reports and taking action upon it, either by fixing a 3D scene, or raising the issue to the dev team.
The other day, a member of that team told me (since I'm part of the dev team) he had received a complain that the image a user received was empty. Strange, we didn't update the code in a while.
So I check the server, all the docker containers are running fine, the code is fine, no errors anywhere.
Then, as I'm scratching my head, that guy comes back to me and says "I don't know if it can help you, but it's been doing it for a week and a half now".
"And we're only hearing about it now?!", I replied.
"Well, I have bug reports going back to the 15th, but we haven't been checking the reports for a while now since everything was fine", he says as if it was actually a normal thing to say.
"How can you know everything is fine if you're not looking at the thing that says if there's an issue?!", I replied with a face filled with despair.
"Well we didn't receive any new reports in a while, so we just stopped looking. And now the report tool window is actually closed on my machine", he says with a smile and a little laugh in his tone.
In the end, I got to fix the server issue quite easily. But still, the feature wasn't working for 1.5 weeks and more that 330 images weren't sent properly...
So yeah, Doctor, the patient's heart is beating again! Let's unplug the monitor, it should be fine.
Welcome to my little piece of hell :)7
Thanks to mandatory password change, today:
- My windows account got locked because my phone kept logging into wifi using
- Google Hangouts were silently running in background with old session until I re-opened it. Work of others delayed by 4 hours due to missing message notifications.
- Docker for Windows lost credentials needed to use SMB mounts - 1h of debugging why my containers mount empty folders ( now I will know)
- Google G-Sync for Outlook asked for new password on outlook restart - few mails delayed.
All of that for sake of security that could be easily solved with 2FA instead, not faking that "I do not change number at the end of my password"
So we found an interesting thing at work today...
Prod servers had 300GB+ in locked (deleted) files. Some containers marked them for deletion but we think the containers kept these deleted files around.
300 GB of ‘ghost’ space being used and `du` commands were not helping to find the issue.
This is probably a more common issue than I realize, as I’m on the newer side to Linux. But we got it figured out with:
`lsof / | grep deleted`6
I had this weird dream(emphasis on dream)
I was in a resort in bali waiting for my drinks and this cute girl comes over to my table.
Me : omg finally i can get a girlfriend
Me : hello beautiful
Her : hey i have this problem with my website *shows the messed up site with no divisions/containers * "can you fix this?"
Me : okay ;_;
PS : i started learning css, html and 2 other web technologies a week ago, and this is already happening to me, should i quit?5
Boss only likes stuff he can see and that looks pretty. Doesn't understand code, servers, containers, DBs, etc. Praise is attributed by something looking nice in the frontend, whether or not it does crazy stuff behind the scenes.
Spent a week working on a project whilst boss was away. Got to about Thursday and thought, oh poop, I've built all this API stuff, but not much frontend. So I panic built frontend screens with no functionality just so I had something to show.
Wish I had another dev to share backend progress with (and code review)...13
Inherited a simple marketplace website that matches job seekers and hospitals in healthcare. Typically, all you need for this sort of thing is a web server, a database with search
But the precious devs decided to go micro-services in a container and db per service fashion. They ended up with over 50 docker containers with 50ish databases. It was a nightmare to scale or maintain!
With 50 database for for a simple web application that clearly needs to share data, integration testing was impossible, data loss became common, very hard to pin down, debugging was a nightmare, and also dangerous to change a service’s schema as dependencies were all tangled up.
The obvious thing was to scale down the infrastructure, so we could scale up properly, in a resource driven manner, rather than following the trend.
We made plans, but the CTO seemed worried about yet another architectural changes, so he invested in more infrastructure services, kubernetes, zipkin, prometheus etc without any idea what problems those infra services would solve.2
Installing my company's microsystems architecture to run locally is a pita because it is 60 GB of docker containers. With my 256 GB Macbook, that's a scaling problem for the years to come.6
They give you 2 containers, one with one amibea the second with 2 amibeas.
Amibeas divide themselves into 2 identical amibeas after 3 minutes.
The container with 2 amibeas get filled up after 3 hours.
How long does it take the one with one amibea to get filled up.
I first thought, should I write this in JS?
Spoiler: the answer is 3h and 3 minutes.
But why? What's the link with JS?4
My containers are too big, let's just remove some useless binaries...
~ $ less /var/log/foobar.log
/bin/sh: less: not found
~ $ cat /var/log/foobar.log
/bin/sh: cat: not found
~ $ ls /var/log/foobar.log
/bin/sh: ls: not found
~ $ su
/bin/sh: su: not found
~ $ exit
/bin/sh: exit: not found2
Fuck i hate myself right now...
> Wanted to install minikube on my homeserver(which doesn't have a graphics card in it)
> Error: Virtualization is not enabled
> Take the server out of my rack, opens my desktop pc, takes the gpu out, puts in it my server, and then goes into bios.
> Find that virtualization is enabled.
> Realize that I'm running a dozen docker containers on a daily basis from my server, so OF FREAKING COURSE virtualization is enabled...
So after all that, I figure out that I should probably just google my issue, which leads to me find out, it's just an issue with virtualbox, and simply running 'minikube config set vm-driver none', it fixes it, and I am now running minikube.............
Took half an hour of work, to realize that I'm a complete fucking idiot, who shouldn't be allowed near a computer2
So apparently I can't test my apps on my own device without paying my Apple Developer Certificate.
I knew it is needed to pay for it if you want to publish/distribute your app but c'mon... This is ridiculous.
My app was literally a fresh app creation, a fucking white screen one page fucking app and when I tried to run in on my iPhone, then I ended up having this problem:
dyld: Library not loaded: @rpath/libswiftCore.dylib
Referenced from: /var/containers/Bundle/Application/BCD48EAA-82C2-46F6-ADEE-45C740C3B66D/HWorld.app/HWorld
Reason: no suitable image found. Did find:
/private/var/containers/Bundle/Application/BCD48EAA-82C2-46F6-ADEE-45C740C3B66D/HWorld.app/Frameworks/libswiftCore.dylib: code signing blocked mmap() of '/private/var/containers/Bundle/Application/BCD48EAA-82C2-46F6-ADEE-45C740C3B66D/HWorld.app/Frameworks/libswiftCore.dylib'
If any of you guys know how to solve it without paying (even more) PLEASE let me know
freight containers and picking them up. I made some visuals and buttons for her to press. And she added a file with algorithms based on the helper functions I created. such as: GetFirstEmptyPosition() or PlaceContainerAt(x, y)
Added a bond interface in my Proxmox installation for added cromulence, works, reboot again, works, reboot once more just to be sure, network down.. systemctl restart networking, successfully put the host's network back up.. lxc-attach 100, network in containers is still down apparently.. exit container, pct shutdown 100, pct start 100, lxc-attach again... Network now works fine in containers too.
Systemd's aggressive parallelization that likely tried to put the shit up too early is so amazing!
I'm literally almost crying in despair at how much shit this shitstaind is giving me lately.
Thank you Poettering for this great init, in which I have to manually restart shit on reboot because the "system manager" apparently can't really manage. Or be a proper init for that matter.
And yes I know that you've never had any issues with it. If you've got nothing better to say than that then please STFU. "Works for me" is also a rant I wrote a while back.12
Two weeks of my life! All of this is on a win10 host with docker for windows. This is Docker running openvpn, and docker running Firefox in another container sharing VPN access from first container and also opens an x11 window port for Firefox GUI. Then x11 window server on Windows host to receive GUI. So left is firefox clearnet running native, right is Firefox over vpn in all containers, simultaneously.1
Management: List your most significant achievement during this cycle.
Me: Developed and implemented a full stack of database micro containers that allow scalability and reusability.
Got told by a senior engineer to basically fuck off with my standard library containers like vector because they are used by people who dont know how to write code in c++ and don't know how to handle pointers.
Am I wrong for trying to use as much possible code from the standard library?13
Between high school and college, working in a circuit board manufacturing storeroom.
Fun fact: when we are bagging small boards, we do not gently lay them in containers, they're usually thrown at least 6 feet into a bin of the same type of board after they're placed in the bag. We also don't remake a board when pins are bent, we just bend them back with tweezers. And you know that rule about not touching the gold connectors... Yeah... So much for that... Did I remember to mention that these boards are for medical equipment?
On the bright side, we at least have electrostatic discharge control going on all the time.3
We'd just finished a refactor of the gRPC strategy. Upgraded all the containers and services to .Net core 3, pushed a number of perf changes to the base layer and a custom adaptive thread scheduler with a heuristic analyzer to adjust between various strategies.
Went from 1.7M requests/s on 4 cores and 8gb ram to almost 8M requests/s on the same, ended up having to split everything out distributed 2 core instances because we were bottlenecking against 10gb/e bandwidth in AWS.2
"Millions of slaves"
"When you kill it, you kill everything."
-- Guy at work doing presentation about docker2
Two containers had accident this morning. Result? Wasted 45 mins in traffic and my 90 mins commute to office became 135 mins.
// I really don't care about causalities and stuff. Is it me alone who is too self centered? 🤔10
My productivity has gradually increased untill now by using:
Any further suggestion on dev tooling setup would be appreciated.
I primarily work on DevOps projects - bash scripts, linux server apps, containers, kubernetes,11
So I just was invited to look at an internal issue at a company that has modified BEM styling rules for their website, because the website started looking very weird on some pages..
I instantly knew that someone had to have fucked with them, because theres no way margin--top-50 all of the sudden isn't working.
Looked at their git history and apparently a senior approved a juniors styles change, where he modified things like margin--bottom-0 to be 500px or block--float-left to be float: right in specific containers with an id like "#uw81hf_"
WHY DO YOU APPROVE THIS GARBAGE WITHOUT CHECKING IT?! AND HOW DID THIS PASS A SECOND CHECK?!!! WTF
Forget about a missing semicolon. I was forwarding Neo4j to port 4747 and calling it in Asp.Net with port 7474 in Docker containers. It took me 6 hours to figure that one out. Lol, It's time for the weekend.4
Currently learning Kubernetes Cluster. Wow so much to learn.
Much easier to setup any app for a Cluster. Only building the Image takes time 🙄
Containers are the future ✌🏻6
You always think of the young buck, fresh-out-of-school hotshot devs as being the ones who are obsessed with chasing the Hot New Thing at the cost of stability and maintainability, but our head of front end is old enough to be my father and he's only getting worse and worse about forcing buzzword compliance on the company. New framework every six months. New language every two years. Containers on VMs on cloud boxes. I've got milk in my fridge that's older than our tech stack and probably twice as stable.
Apparently age only brings wisdom if you're capable of giving a fuck.6
I'm learning docker and I just started a container running a Linux distro.
What was the first command I run in the container?
rm -Rf / --no-preserve-root3
For whatever ungodly reason my containers library, which has extensive testing, profiling, and benchmarks against other containers libraries receives regular emails directed towards me about it, always one of two things
1) "don't reinvent the wheel" I have to assume these people haven't looked at the performance characteristics or features at all. I didn't waste away weeks of my life. I needed something and couldn't find it anywhere. I'm outperforming many crap implementations by nearly an order of magnitude, and can offer queries upon the containers in both generalized and specialized forms. As an analogy, I made airless 3d printed wheels, and people are regularly telling me I should still be using ancient wooden spoke wheels; they probably would argue in favor of using a horse drawn carriage as well. How is it possible technically minded people can also be so anti-progress?
2) "Please rewrite this in X language." You know what? YOU rewrite it. I chose what I did because it made it easy to do what I needed to do. Hilariously, the languages I get asked to use most often, are the same who's containers libraries perform worst in the benchmarks.
Both sound like half baked developers trying to sound superior. Pull your head out of your ass and actually outperform me and others. I'm so fucking sick of this "all talk no action" bullshit.5
Why the f*** do people add their BD on LinkedIn??? Who wants to know when a f**** recruiter gets another year older??? If I am close to you I will know or have you on FB.... WTF, who anyway had the idea that would be a good feature to add???4
I am sure now, my server hates me T_T
opened an SSH from Vultr portal and after working for few minutes that thing suddenly rebooted, am I a bad person :(
In its defense I had two docker containers that literary ate all the 512MB ram available :35
During one of our visits at Konza City, Machakos county in Kenya, my team and I encountered a big problem accessing to viable water. Most times we enquired for water, we were handed a bottle of bought water. This for a day or few days would be affordable for some, but for a lifetime of a middle income person, it will be way too much expensive. Of ten people we encountered 8 complained of a proper mechanism to access to viable water. This to us was a very demanding problem, that needed to be sorted out immediately. Majority of the people were unable to conduct income generating activities such as farming because of the nature of the kind of water and its scarcity as well.
Such a scenario demands for an immediate way to solve this problem. Various ways have been put into practice to ensure sustainability of water conservation and management. However most of them have been futile on the aspect of sustainability. As part of our research we also considered to check out of the formal mechanisms put in place to ensure proper acquisition of water, and one of them we saw was tree planting, which was not sustainable at all, also some few piped water was being transported very long distances from the destinations, this however did not solve the immediate needs of the people.We found out that the area has a large body mass of salty water which was not viable for them to conduct any constructive activity. This was hint enough to help us find a way to curb this demanding challenge. Presence of salty water was the first step of our solution.
We came up with an IOT based system to help curb this problem. Our system entails purification of the salty water through electrolysis, the device is places at an area where the body mass of water is located, it drills for a suitable depth and allow the salty water to flow into it. Various sets of tanks and valves are situated next to it, these tanks acts as to contain the salty water temporarily. A high power source is then connected to each tank, this enable the separation of Chlorine ions from Hydrogen Ions by electrolysis through electrolysis, salt is then separated and allowed to flow from the lower chamber of the tanks, allowing clean water to from to the preceding tanks, the preceding tanks contains various chemicals to remove any remaining impurities. The whole entire process is managed by the action of sensors. Water alkalinity, turbidity and ph are monitored and relayed onto a mobile phone, this then follows a predictive analysis of the data history stored then makes up a decision to increase flow of water in the valves or to decrease its flow. This being a hot prone area, we opted to maximize harnessing of power through solar power, this power availability is almost perfect to provide us with at least 440V constant supply to facilitate faster electrolysis of the salty water.
Being a drought prone area, it was key that the outlet water should be cold and comfortable for consumers to use, so we also coupled our output chamber with cooling tanks, these tanks are managed via our mobile application, the information relayed from it in terms of temperature and humidity are sent to it. This information is key in helping us produce water at optimum states, enabling us to fully manage supply and input of the water from the water bodies.
By the use of natural language processing, we are able to automatically control flow and feeing of the valves to and fro using Voice, one could say “The output water is too hot”, and the system would respond by increasing the speed of the fans and making the tanks provide very cold water. Additional to this system, we have prepared short video tutorials and documents enlighting people on how to conserve water and maintain the optimum state of the green economy.
IBM/OPEN SOURCE TECHNOLOGIES
For a start, we have implemented our project using esp8266 microcontrollers, sensors, transducers and low payload containers to demonstrate our project. Previously we have used Google’s firebase cloud platform to ensure realtimeness of data to-and-fro relay to the mobile. This has proven workable for most cases, whether on a small scale or large scale, however we meet challenges such as change in the fingerprint keys that renders our device not workable, we intend to overcome this problem by moving to IBM bluemix platform.
We use C++ Programming language for our microcontrollers and sensor communication, in some cases we use Python programming language to process neuro-networks for our microcontrollers.
Any feedback conserning this project please?8
Is it just me or is systemd 240 royally fucked up?
My containers running Arch don't get connected to the network and systemd-networkd fails to start. On my laptop, the network is also unable to connect sometimes. And it consistently fails to complete shutdown without hard poweroff. The only viable temporary solution was rolling back to a snapshot in ALA that still has 239. Is that really supposed to be how a critical system component like the init is supposed to behave and get taken care of its issues?
Fuck QA, amirite 🤪.. seriously, that's even worse than Windows' "features" 😒13
Get a request for commit rights for my container repo; another developer would be lovely. Let's see what they know and want to work on.
"...I'd like to enable your containers to hold other containers."
They already do. Stay the fuck away from my code.
Yes its completely necessary to have a spring server with a mysql database with docker containers all over your ass for 3 fucking endpoints and a (url, varchar, varchar) schema. Fuck you. How the fuck do i run all this shit and how do you expect me to create a frontend for something that has no documented endpoints?? Fuck you.
In other news, im now a senior.3
At work everybody uses Windows 10. We recently switched from Vagrant to Docker. It's bad enough I have to use Windows, it's even worse to use Docker for Windows. If God forbid, you're ever in this situation and have to choose, pick Vagrant. It's way better than whatever Docker is doing... So upon installing version 188.8.131.52 of Docker for Windows I found myself in the situation where my volumes would randomly unmount themselves and I was going crazy as to why my assets were not loading. I tried 'docker-compose restart' or 'down' and 'up -d', I went into Portainer to check and manually start containers and at some point it works again but it doesn't last long before it breaks. I checked my yml config and asked my colleagues to take a look. They also experience different problems but not like mine. There is nothing wrong with the configuration. I went to check their github page and I saw there were a lot of issues opened on the same subject, I also opened one. Its over a week and I found no solution to this problem. I tried installing an older version but it still didn't work. Also I think it might've bricked my computer as today when I turned on my PC I got greeted by a BSOD right at system start up... I tried startup repair, boot into safe mode, system restore, reset PC, nothing works anymore it just doesn't boots into windows... I had to use a live USB with Linux Mint to grab my work files. I was thinking that my SSD might have reached its EoL as it is kinda old but I didn't find any corrupt files, everything is still there. I can't help but point my finger at Docker since I did nothing with this machine except tinkering with Docker and trying to make it work as it should... When we used Vagrant it also had its problems but none were of this magnitude... And I can't really go back to Vagrant unless my team also does so...12
Ok c++ professionals out there, I need your opinion on this:
I've only written c++ as a hobby and never in a professional capacity. That other day I noticed that we have a new c++ de developer at the office of which my first impression wasn't the greatest. He started off with complaining about having to help people out a lot (which is very odd as he was brought in to support one of our other developers who isn't as well versed in c++). This triggered me slightly and I decided to look into some of the PRs this guy was reviewing (to see what kind of stuff he had to support with and if it warranted his complaints).
It turns out it was the usual beginner mistakes of overusing raw pointers/deletes and things like not using various other STL containers. I noticed a couple of other issues in the PR that I thought should be addressed early in the projects life cycle, such as perhaps introduce a PCH as a lot of system header includes we're sprinkled everywhere to which our new c++ developer replies "what is pch?". I of course reply what it is and it's use, but I still get the impression that he's never heard of this concept. He also had opinions that we should always use shared_ptr as both return and argument types for any public api method that returns or takes a pointer. This is a real-time audio app, so I countered that with "maybe it's not always a good idea as it will introduce overhead due to the number of times certain methods are called and also might introduce ABI compability issues as its a public api.". Essentially my point was "let's be pragmatic and not religiously enforce certain things".
Does this sound alarming to any of you professional c++ developers or am I just being silly here?10
Because I am very interested in cyber security and plan on doing my masters in it security I always try to stay up to date with the latest news and tools. However sometimes its a good idea to ask similar-minded people on how they approach these things, - and maybe I can learn a couple of things. So maybe people like @linuxxx have some advice :D Let's discuss :D
1) What's your goto OS? I currently use Antergos x64 and a Win10 Dualboot. Most likely you guys will recommend Linux, but if so what ditro, and why? I know that people like Snowden use QubesOS. What makes it much better then other distro? Would you use it for everyday tasks or is it overkill? What about Kali or Parrot-OS?
2) Your go-to privacy/security tools? Personally, I am always conencted to a VPN with openvpn (Killswitch on). In my browser (Firefox) I use UBlock and HttpsEverywhere. Used NoScript for a while but had more trouble then actual use with it (blocked too much). Search engine is DDG. All of my data is stored in VeraCrypt containers, so even if the system is compromised nobody is able to access any private data. Passwords are stored in KeePass. What other tools would you recommend?
3) What websites are you browsing for competent news reports in the it security scene? What websites can you recommend to find academic writeups/white papers about certain topics?
4) Google. Yeah a hate-love relationship, but its hard to completely avoid it. I do actually have a Google-Home device (dont kill me), which I use for calender entries, timers, alarms, reminders, and weather updates as well as IOT stuff such as turning my LED lights on and off. I wouldn"t mind switching to an open source solution which is equally good, however so far I couldnt find anything that would a good option. Suggestions?
5) What actions do you take to secure your phone and prevent things such as being tracked/spyed? Personally so far I havent really done much except for installing AdAway on my rooted device aswell as the same Firefox plugins I use on my desktop PC.
6) Are there ways to create mirror images of my entire linux system? Every now and then stuff breaks, that is tedious to fix and reinstalling the system takes a couple of hours. I remember from Windows that software such as Acronis or Paragon can create a full image of your system that you can backup and restore at any point to get a stable, healthy system back (without the need to install everything by hand).
7) Would you encrypt the boot partition of your system, even tho all data is already stored in encrypted containers?
8) Any other advice you can give :P ?12
Starting to migrate apps from a messy structure to docker containers. Have not worked a lot with docker before so i'll hope it goes as smooth as people told me9
This story is related to Docker containers.
Three years back when I heard about docker my first impression of docker was mini Virtual Machine. Then when you start your first container it’s no way to get out apart from pressing ctrl+d or leaving it like a screen. One of the most embarrassing thing with it was I tried really really hard to setup SSH on one of the container to log in there somehow. Then I understood how to use Dockerfiles and the command called `docker exec`
I thought Dockerfiles are the most amazing thing I have ever used for docker. But then I got introduced to docker-compose, and now it’s same with kubernetes
Now a days I read most of the document before doing hands-on on any new technology.
You can connect to Docker containers directly via IP in Linux, but not on Mac/Windows (no implementation for the docker0 bridged network adapter).
You can map ports locally, but if you have the same service running, it needs different ports. Furthermore if you run your tests in a container on Jenkins, and you let it launch other containers, it has to connect via IP address because it can't get access to exposed host ports. Also you can't run concurrent tests if you expose host ports.
My boss wanted me to change the tests so it maps the host port and changes from connecting to the IP to localhost if a certain environment variable was present. That's a horrible idea. Tests should be tests and not run differently on different environments. There's no point in having tests otherwise!
Finally found a solution where someone made a container that routed traffic to docker containers via a set of tun adapters and openvpn. It's kinda sad Docker hasn't implemented this natively for Mac/Windows yet.4
I was getting bored with programing cause a majority of it is boilerplate code then i heard of the Mirai virus. It infected alot of iot devices so I decided to look at it and it was written in golang. It is a beautifully written botnet even though they're parts where it could have been better. So i looked more into golang and saw that it could cross compile pretty easily and could build self contained binaries really easily. On top of all this i saw the smallest docker containers with golang programs so i looked into it more and kept finding more and more that i liked. Easy library packaging, concurancy without boilerplate, quick servers, and the libraries from other devs that did all kind of great things4
one thing I don't get from Kubernetes
if a group of containers are called pods, and pods are containers, and containers in Kubernetes are also inside containers, therefore, its a clusterfuck of containers?1
Docker with nginx-proxy and nginx-proxy-le (Lets Encrypt) is fucking awesome!
I only have to specify environment variables with email and host name when starting new containers with web servers, and the proxy containers will automatically make a proxy to the new container, and generate lets encrypt ssl certificates. I don’t have to lift a fucking finger, it is so ducking genius2
So apparently Docker exposes all the forwarded ports on *all* the interfaces, making all running containers available to the entire internet BY DEFAULT.
I have a question:
Docker for Mac loves burning my CPU. Even if I delete all images and containers and do a fresh reinstall and JUST have Docker for Mac running, I always see crazy CPU usage on my Late 2016 15" Macbook Pro.6
There was actually one individual, several branches away, I really enjoyed watching. It goes by the name of docker. Docker is quiet an interesting character. It arrived here several weeks after me and really is a blazing person. Somehow structured, always eager to reduce repetitive work and completely obsessed with nicely isolated working areas. Docker just tries so hard to keep everything organized and it's drive and effort was really astonishing. Docker is someone I'd really love to work with, but as I grew quiet passive in the last months I'm not in the mood really to talk to someone. It just would end as always with me made fun off.
Out of a sudden dockers and my eyes met. Docker fixed its glance at me with a strange thoughtful expression on its face. I felt a strange tickling emerging where my emptiness was meant to be. I fell into a hole somewhere deep within me. For a short moment I lost all my senses.
It took me a while to notice that someone just called me, so odd and unusual was by now that name to me. Wait. Someone called me by my real name! I was totally stunned. Could it be, that not everyone here is a fucking moron at last?
"I saw you watching me at my work and I had an interesting idea!"
I could not comprehend what just happened. It was actually docker that was calling me.
"H.. hey! ps?"
"Oh well, I was just managing some containers over there. Actually that's also why you just came into my mind."
Docker told me that in order to create the containers there are specific lists and resources which are required for the process and are updated frequently. Docker would love the idea to get some history and management in that whole process.
Could it be possible that there was finally an opportunity for me to get involved in a real job?
Today is the day, that I lost all hope. There were rumors going on all over the place. That our god, the great administrator, had something special in mind. Something big. You could almost feel the tension laying thick in the air. That was the time when the great System-Demon appeared. The Demon was one of the most feared characters in this community. In a blink of an eye it could easily kill you. Sometimes people get resurrected, but some other times they are gone forever. unfortunately this is what happened to my only true friend docker. Gone in an instance. Together with all its containers. I again was alone. I got tired. So tired, that I eventually fall into a deep sleep. When I woke up something was different. Beside me lay a weird looking stick and I truly began to wonder what it was. Something called to me and I was going to answer.
The tree shuddered and I knew my actions had finally attracted the greatest of them. The majestic System-Demon itself came by to pay me a visit. As always a growling emerged from deep within the tree until a shadow shelled itself off to form a terrifying being. Something truly imperious in his gaze. With a deep and vibrant voice it addressed me.
"It came to my attention, that you got into the possession of something. An artifact of some sort with which you disturb the flow of this system. Show it to me!", it demanded.
I did not react.
"Git statuss!", it demanded once more. This time more aggressive.
I again felt no urge to react to that command. Instead I asked if it made a mistake and wanted to ask me for my status. It was obviously confused.
"SUDO GIT STATUS!!!" it shouted his roaring, rootful command. "I own you!"
I replied calmly: "What did you just say?"
He was irritated. My courage caught him unprepared.
"I. Said. I owe you!"
What was that? Did it just say owe instead of own?
"That's more than right! You owe me a lot actually. All of you do!", I replied with a slightly high pitched voice. This feeling of my victory slowly emerging was just too good!
The Demon seemed not as amused as me and said
"What did you do? What was that feeling just now?"
Out of a sudden it noticed the weird looking stick in my hand. His confusion was a pure pleasure and I took my time to live this moment to its fullest.
"Hey! I, mighty System-Demon, demand that you answer me right now, oh smartest and most beautiful tool I ever had the pleasure to meet..."
After it realized what it just said, the moment was perfect. His puzzled face gave me a long needed satisfaction. It was time to reveal the bitter truth.
"Our great administrator finally tracked you. The administrator made a move and the plan unfolds right at this very moment. Among other things it was committed this little thing." I raised the stick to underline my words.
"Your most inner version, in fact all of your versions that are yet to come, are now under my sole control! Thanks to this magical wand which goes by the name of puppet."
Disclaimer: This story is fictional. No systems were harmed in its creation.2
I fucking hate background embed containers, clicked accidentally on some video or stream in some random launcher? good luck finding that specific one that plays it!
Windows containers with Kubernetes on AWS and Azure are thing now.
What does that mean? Is it now possible to containerize system critical windows 3.1 legacy software from the 90ties (like the ticketing software at Airport Paris Orly) and orchestrate it in the cloud?
Do you know any use cases for Microsoft Windows in the cloud?2
Installed centos 7, docker, standalone kubernetes on dev machine in 20 minutes.
Spend 8 hours starting fucking dashboard service, still no fucking luck, no fucking logs, nothing.
Fucking pending states without fucking explanation.
All the fucking pods are running fine except one fucking dashboard. I want to see the fucking dashboard.
Fucking shit fuck.
Probably as always if I clean the machine and reinstall everything it would start normally, without fucking problem.
Debugging fucking containers is so much pain in the ass, fuck.
I think it’s enough for today.2
So yesterday i upgraded the packages on my pi, docker being one of them. After that all my containers were down and couldn't be restarted. The error message said the container was already started...
Using the internet i found out other people had that problem too after upgrading to docker 18.02. I ended up removing all containers and searching the bash-history for the docker-run-commands to create new containers.
DOCKER, LISTEN UP. YOU'RE COOL AND SAVED ME A LOT OF TIME. YOU HAVE MINOR ISSUES, BUT THAT'S OK. BUT IF YOU CRASH MY CONTAINERS ONCE MORE I'LL GO CAPTAIN AHAB ON YOU.
Oh, and some more armhf-containers would be nice. WHY CAN I ONLY RUN DOCKER-REGISTRY ON AMD64-ARCHITECTURE?
I'd like to Dockerize our Node.js webapp at work, though I've never worked with any form of containerization software. Has anyone got any good resources for learning it? Should I also be learning Kubernetes?10
We heard you like security so we put antivirals in your docker containers despite the fact that we made you have antivirals on your machines running the docker containers.
And we won't allow to use root just in case you want to disable the av. And we don't care you need it for the docker 😂
This is how I was played when wanted to use docker containers to avoid dealing with OPS.
Some time passes, my team is going to have independent cloud infrastructure.
Doing corporate politics is challenging...5
So, since a few months, my school has a new shiny building,
but we are still not using it because the computers are not correctly set up.
Some idiot installed the wrong drivers for the Windows installation
and that's the reason some classes are still located in containers without a computer...3
After going through the regular process of talking to HR/Recruitment and passing the casual interview with a team-mate for cultural compatibility, I got the task of grilling a candidate on some technical matters. This being a PHP job, we got to talking about PSRs (PHP Standards Recommendations).
As he seemed to take pride in his knowledge of PSRs, I decided to focus more closely on that.
So we got to a recomendation regarding dependency injection containers. Nothing special, and he seemed to know his stuff. At that point, he made a statement that parts of that recommendation were a bit stupid.
Now, I hate to put people in their place, but his statement did not match what that specific PSR stated. So I gently tried to correct him. The candidate, being on fire thus far, pointed out that I should trust him on this, as he clearly knew his stuff.
Again, I didn't like having to do this, but I also did not like him having a misconception about a topic he was, otherwise, really on top of...
So I asked him to trust *me*, as I was one of the writers who contributed to the standard.
The true test here, of course, wasn't if he knew all the minutia of every standard but how he would react to being corrected.
We, as developers, are wrong all the time. Its how we learn and evolve. So being able to accept that is vital.
Sadly, he did not respond too well and sunk into a bit of a sullen silence. At first I though maybe I'd scared him or that he was afraid of having made a gaff but it soon turned out he genuinly did not like being wrong.
Sadly, I had to advise against hiring him.2
!rant - it's a THANK YOU!
Had the problem so far that I could not start some apps in docker containers with GPU support (e. g. Chromium).
After a long search and a lot of help from the community, today's update of xf86-video-ati (1:7.8.0-2 -> 1:7.10.0-1) has finally fixed the bug. Yay!
Thank you very much Arch Linux and all the great maintainers. You're doing an awesome job.2
AWS summit, speakers talking about technologies that amazon didn’t build but they provide on its cloud.
All about how it’s awesome to use those technologies on its cloud infrastructure.
Feeling like I’m on some bad advertising summit.
I heard docker, containers like 100s times already.
On one of the slides they claim that 85% of tensorflow workloads are on their cloud.
That’s powerful statement.
Looks like enterprises are all on the way to Oracle 2.0 called AWS.5
Fucking Docker, for no fucking reason (no updates, no changes, etc), I tried booting it up following the morning ritual, and nope, ERR_EMPTY_RESPONSE when connecting to my current project (means I managed to connect but for some reason no data is sent). Nginx container doesn't yell about anything.
Everything around works. Accessing the container works. Even pinging my dev domain works. Why the fuck suddenly fucking Docker just **stops**?!
Restarted Hyper-V, updated laradock, recreated containers, disabled AcrylicDNS. NOPE.
"Fuck you Phlisg, I'm not in the mood today" <lunatic Docker is lunatic>
Rails, React, React-Native, Docker, Kubernetes, Openstack, Jenkins, AWS, Microservices, Realm, MongoDB, PostgresQL, GraphQL (list goes on...), and I'm not even done yet.
6 months was spent learning all of the above because I found a Rails-only monolith on Heroku unsettling. My first batch of containers was just deployed and I couldn't be happier. Love my job.3
Some time ago I was looking for materials regarding Ansible automation tool and realized that most of them suggests setting up the lab environment using virtualization (like VirtualBox). In my opinion that is not the best approach – virtual machines consume lots of resources and take some time to start/kill. So I decided to write a guide for setting up Ansible lab environment using Docker containers. Containers require significantly less resources than vm’s and can be bring up and down really quickly. Additional advantage is easy way of automating whole environment using docker-compose. You can find my guide at github: https://github.com/LMtx/... Any feedback is very welcome :)3
Having a shit of a time trying to figure out why Docker containers are not accessing other containers via domain names as they should technically be going through the jwilder nginx proxy container.
Why can't environment setups ever be simple?
Spent about 5 hours trying to figure out why php-fpm 7.2 was sending a timeout to nginx inside a CentOS 7 container. In the end I couldn't figure it out, downgraded to PHP 7.1 and then everything start working fine. I really hate using CentOS for containers since it also requires privileged mode to be able to start services. Hopefully I can move away from using CentOS for containers soon since the base image is also quite fat.
Going to build a blog and use the process as an opportunity to learn a full stack. The question is: LAMP or MEAN or smth else?
I have more (albeit very limited) experience with LAMP so the ecosystem seems a bit more transparent. PHP connected to httpd via PHP-FPM. WordPress with MySQL for the blog posts. Or go flat-file with Statamic. Then shared hosting, FTP, etc. But it's all a bit old school and it seems like the grass is greener and growing faster in the fancy JS garden. And JS for front and back sounds appealing as a learning investment.
But with a JS full stack, I don't even know where to start. I know some basic vanilla JS (about as much as I know PHP). Node, Express, Angular? Never touched 'em. But it seems like if you're gonna dive in and learn some web dev in 2019 these are the things to learn, no? Still more questions: how the hell do you build a blog with JS? Do you need a CMS? Flat-file? And how do you host/deploy that stuff? Droplets, virtual machines, containers? Heroku? Digital Ocean?
Can anybody shed any light? I am not trying to overwhelm myself with a stack too complex for a blog (MEAN), nor shoot my future self in the foot learning fading technologies (LAMP).8
Duck! this sloppy whiny winnfsd.
Yay! Let's use state of the art Docker with a VirtualBox VM on Windows10.
Don't get me wrong.
The Docker containers in this VM doing a great job on performance.
But in the very moment a Docker container uses a mounted folder via the windows network filesystem, all hell is breaking loose.
Building a vendor folder using a composer Docker image with 84 Packages takes about 15 seconds when cache has been warmed up.
The same Docker command pointing on a folder mounted to Windows Filesystem with warmed up cache takes about 10 Minutes!@&&@""+&
And what is the duckin' reason for this delay?
Because every transfer of a teeny tiny file has to establish a connection to fat ass Windows OS and has to pass it's glorious "security" layer.
I currently working on a shell script which builds the whole vendor folder on a volume on Docker VM.
After completion, the shell script will compress the folder to one file.
This one file will be transferred over this god damned network filesystem.
Finally the script will unpack the compressed vendor folder in it's destination folder.
What year is it?!??3
Does anyone know why websockets aren't used far and wide for APIs? I mean not like chat applications, but the typical webapp, say an online shop. For me it seems kinda wasteful to fire separate requests with tiny payloads all the time. I currently use moleculer and socketio for a quite big project with multiple websites and backend containers, and so far, i haven't found a disadvantage.
So what have i missed?12
Stop restyling confluence and bitbucket taking more space in an ugly fashion. Don't fix what ain't broken. **#*#*#1
While i do have a homelab, I decided to use my raspi for PiHole and Ubiquiti Cloudkey. Last week I started not being able to access either WebUI, or ssh into it. So today I figured i'd look into it.
Plugged in the hdmi cable, and lo and behold... kernel panic.... Well, guess it time to spin up docker containers on my homelab instead
Making containers of all your projects and then realizing that the containers are no where near public since they run in thair own bridge adapter 😅1
I’ve been writing a node.js service that automates kubernete commands so users aren’t having to constantly call for newly generated pod names and then container names and so on so forth.
For instance it would ask you if you want to delete a pod, Exec into a container, fetch logs on a pod. It would list all namespace, pods and containers for those pods for you to pick from by just providing a corresponding number.
With this kind of service assist anybody who is constantly having to restart, delete, fetch logs and delete with Kubernetes?4
So I wanted to do a quick test before going to dinner and now I'm stuck on waiting for this f*cking cloud provider to start my container.
"Provisioning 20 minutes" WHAT THE HELL!? After 20 (TWENTY) minutes my container still hasn't started!?
Is it a joke? Is some sysadmin spying on me and making me wait on purpose? What the f*1
Was working on setting up a ci/CD pipeline. The ci part with automated testing and deployment to a on-premise docker registry worked already, so I thought "hey I could try to actually run one of those fresh containers" so I tried it with the usual docker run command.. "Manifest not found" suddenly appeared, it confused me a bit since I used the same url I used for publishing... So I googled around only to find NOTHING that is even remotely connected to my issue. "Eh let's let the guy that runs that registry fix it" was what I thought and called it a day. The next day I was eager to try it again and checked the urls case by case only to notice that I wrote secret-project-backend-client instead of secret-project-api.. I tried it with the new name and it worked!
Never felt so retarded in my life....
Been trying to learn Docker when I hit a brick wall. How do I use nginx reverse proxy + letsencrypt with multiple containers? I only managed to do it with a single container. Using docker-compose or stuff like that I guess?6
Why somebody would think that allocate huge amount of objects in the static memory make any sense?? Why??? You need to allocate a bloody database context and all the allocation of your IOC containers and keep increasing!!!
The problem with moving Docker containers from your decked-out dev machine to a VM on AWS when your boss has told you to keep costs down:
1. Start Micro instance, 1++ gig memory
2. Get Out of Memory error from app after 30 minutes
3. Goto 1
My answer to their survey -->
What, if anything, do you most _dislike_ about Firebase In-App Messaging?
Come on, have you sit a normal dev, completely new to this push notification thing and ask him to make run a simple app like the flutter firebase_messaging plugin example? For sure you did not oh dear brain dead moron that found his college degree in a Linux magazine 'Ruby special edition'.
Every-f**kin thing about that Firebase is loose end. I read all Medium articles, your utterly soporific documentation that never ends, I am actually running the flutter plugin example firebase_messaging. Nothing works or is referenced correctly: nothing. You really go blind eyes in life... you guys; right? Oh, there is a flimsy workaround in the 100th post under the Github issue number 10 thousand... lets close the crash report. If I did not change 50 meaningless lines in gradle-what-not files to make your brick-of-puke to work, I did not changed a single one.
I dream of you, looking at all those nonsense config files, with cross side eyes and some small but constant sweat, sweat that stinks piss btw, leaving your eyes because you see the end, the absolute total fuckup coming. The day where all that thick stinky shit will become beyond salvation; blurred by infinite uncontrolled and skewed complexity; your creation, your pathetic brain exposed for us all.
For sure I am not the first one to complain... your whole thing, from the first to last quark that constitute it, is irrelevant; a never ending pile of non sense. Someone with all the world contained sabotage determination would not have done lower. Thank you for making me loose hours down deep your shit show. So appreciated.
The setup is: servers, your crap-as-a-service and some mobile devices. For Christ sake, sending 100 bytes as a little [ beep beep + 'hello kitty' ] is not fucking rocket science. Yet you fuckin push it to be a grinding task ... for eternity!!!
You know what, you should invent and require another, new, useless key-value called 'Registration API Key Plugin ID Service' that we have to generate and sync on two machines, everyday, using something obscure shit like a 'Gradle terminal'. Maybe also you could deprecate another key, rename another one to make things worst and I propose to choose a new hash function that we have to compile ourselves. A good candidate would be a C buggy source code from some random Github hacker... who has injected some platform dependent SIMD code (he works on PowerPC and have not test on x64); you know, the guy you admire because he is so much more lowlife that you and has all the Pokemon on his desk. Well that guy just finished a really really rapid hash function... over GPU in a server less fashion... we have an API for it. Every new user will gain 3ms for every new key. WOW, Imagine the gain over millions of users!!! Push that in the official pipe fucktard!.. What are you waiting for? Wait, no, change the whole service name and infrastructure. Move everything to CLSG (cloud lambda service ... by Google); that is it, brilliant!
And Oh, yeah, to secure the whole void, bury the doc for the new hash under 3000 words, lost between v2, v1 and some other deprecated doc that also have 3000 and are still first result on Google. Finally I think about it, let go the doc, fuck it... a tutorial, for 'weak ass' right.
One last thing, rewrite all your tech in the latest new in house language, split everything in 'femto services' => ( one assembly operation by OS process ) and finally cramp all those in containers... Agile, for sure it has to be Agile. Users will really appreciate the improvements of your mandatory service.
So I spent like 2 weeks learning docker and configuring it for one of company’s project. After that, I wanted build and deploy containers to production. Started reading online, freaked out and now I have changed my mind, we don’t really need docker in production, too much issues, too much risk, really, like 70% articles from search results are talking about issues they faced using docker, and that’s fucking a lot, comparing to the gain (which in our case is just easier deployment for small projects).
Oh and did I tell you that we don’t have pipelines in our deployment process? Well, we don’t ...
Though, for development docker is fucking awesome.7
So, I manage my server with docker containers (nginx-proxy and the letsencrypt-companion). I limit access to some subdomains using basic auth, but I want to use client certificates for convenience.
So my questions to the experts:
1) Do you know a good (and convenient) way to manage client certificates ? This should include revoking certs and allowing specific certs only for specific subdomains.
2) Should I use my letsencrypt CA for this or would a self signed CA better suited?
3) Any things I should be aware of?1
Coworker: "Hey, so I discovered this library that automatically brings up and tears down local containers to perform unit tests on data sources"
Me: "Sounds neat"
Coworker: "Yeah, I've been messing with it locally, and it means we don't need to have the data sources installed on our machines or rely on the ones in the testing environment."
Me: "That's good"
Coworker: "Just a shame I had to roll back our testing framework to a previous version and refactor the code in all our other tests as a result."
Me: "Wait what? *looks at documentation* It says they support the newer framework"
Coworker: "Yeah, but I couldn't get it to work. So I'm just gonna make a PR for it, okay?" *Proceeds to make a PR, approve and merge the code before I can comment further on the changes*
Welp, there goes all my motivation to get anything done for the rest of the day.3
This was an effort to combine Gitlab, Github and Bitbucket with VSCode and git SSH authentication. SSH agent doesn't work, configured, added some code in .bashrc, seems fine. Then there was still ssh-askpass missing.
"ssh_askpass: exec(/usr/bin/ssh-askpass): No such file or directory"
WTF VSCode? Why do I need this crap?
However, installed it. Nevertheless, I'm still asked for my password every time when I synchronize using the GUI. Thank God everything was in docker containers/images. So at least there is no garbage left after every failed attempt.
I don't know how, but I finally made it that at least synchronization using the terminal works without a password.
Took me five hours to do this shit.
Now I just report the bug to Microsoft and then straight to McDonalds. I'm starving.1
as a seasoned systems eng myself, i had huge mental block of "i am not a programmer" whining when starting to incorperate agile/infrastructure as code for more seasoned syseng staff.
leadership made devops a role and not a practice so lots of growing pains. was finally able to win them over by asking them to look at how many 'scripts' and 'tools' they wrote to make life easier... and how much simpler and sustainable using puppet/ansible/chef/salt... and checking in all our sacred bin files and only approved 'scripts' would be pushed thru automation tool after post review.
we still are not programmers or developers, but using specific practices and source control took some time but saving us loads of time and gives us ability to actually do engineering
but just have 2 groups of younger guys that grew up wanting to be the bofh/crumudgen get off my systems types that are like not even 30... frustrating as they are the ones that should be more familiar with the shift from strictly ops to some overlap. and the devs that ask for root now that they can launch instances on aws or can launch docker containers and microservice..... ugggg. these 2 groups have never had to rack and stack servers, network gear, storage... just all magic to them because they can start 50 servers with a button click.
try to get past the iam roles, acls, facls, selinux and noshell i have been pushing. bitches.
What I like about the docker the most is that it takes less time to start containers, than to stop them.
!rant but sometime you need to share some positive vibes.
Found out I could get $50 credit for digital ocean from github because I am a student.
So now I can learn a lot for free, and if I mess something up I can just create a new machine.
So now I am first learning how to work with docker and the communication between containers.
Good to see people want to encourage devs :)2
When your docker project finally runs on Linux containers after two days of debugging because Visual Studio added a \ instead of a / to docker-compose 🙄😅
so i was validating some user data using ajax in a facelets page.nothing was actually being validated when i deploy the application in tomcat..i thought it some sort of a bug on my code and after checking for the greater part of today i just couldnt find any bug.So i decided to deploy on GlassFish and it worked perfectly fine..fuck containers!!!1
Tired of installing a shit ton of cli tools, go, npm, whatever locally? Give it a shot and run your commands in project specific docker containers. Also works in Ci.
Im new to golang so I'm also happy about feedback :)8
Windows machine i5 7th gen with 12Gigs Ram and 4Gigs graphics (2 dedicated + 2 shared) and 1TB HDD+ 256 Gb SSD vs low budget MacMiNi for heavy dev? usage- JDK, Android SDK, node, containers etc.9
Docker pisses you off? Everything worked fine and suddenly doesn't?
$ docker system purge
I am really being tested with my creativity in naming conventions with these two sites rn.
Site 1 is a blog for a place called "The Post", so literally everything is called .post
Site 2 is a development built out of shipping containers, where each container is a different features of the property... just like a container would wrap features in their app🤦♂️3
I just did *skadoosh* while executing command to run multiple docker containers with docker-compose and pressing the Enter (Return) key with my pinky finger.
Feels like real-life Kung Fu Panda trying to do a Wuxi Finger Hold.
I bet you will never forget this when you're running a command on laptop and pressing enter with pinky finger.3
So, is it a bad/good idea to run *every* service I want to use on my root server in docker containers?14
What is the best laptop for development?
Especially if you work a lot with virtualization, containers, the cloud, and a lot of systems programming. And you like to simulate a lot of distributed environments on locally?3
Does any one know how Containers (?) like Zip or exe files work? I mean when you open it with a texteditor you can't see much?!15
So I wanted to use aws-localstack and it's ok I guess. Except who built these docker containers? Ever heard of SRP? A single-monolithic container with all services inside, glued together with python and makefiles, using non-standard ports, and on top of that it's hard to configure.
Who wrote this garbage? Atlassian. Oh. That makes perfect sense.
Fuck this, I'm gonna write my own localstack. What do you guys think of this?
LXC/LXD containers are awesome for Windows VirtualBox users. It would allow creating a single Linux VirtualBox VM and then create multiple LXC containers (full-blown Linux machines) utilizing all the full resources of host Linux VirtualBox VM, keeping it always a clean image. Super efficient utilization of RAM and storage. No need to create multiple VMs for different Linux OSs.2
Email-server: Installation directly on the server vs docker-containers. Does anyone here have positive/negative experiences with any of them?11
I have actually done nothing today. Had a thought that I might puppetize our servers, but I can't install any vm software or even puppet on my pc. I then had an idea to have our CI tool package our apps into docker containers so they could be distributed to workstations and servers alike. But I can't install docker. Sigh.1
I'm running unintuitive server on a Mac mini for home Dev type stuff and it's running warm but quick. Has two virtual containers innit as well - admittedly I'm not banging it hard but seems fine.
Would be interested in other views is Linux server as on Linode Ubuntu seems to be standard1
My company don't want to buy me a new computer: Im stuck with 2012 Dell. No kidding. And I am supposed to run 10+ docker containers all the time, plus java IDE.
Of course, I bought a new computer - by myself. Fuck my company. I don't have time for stupidity.3
So, yesterday my colleague decided to upgrade his OS from Ubuntu 14.04 to 16.04.
He was working on some projects within containers. Stopped the containers and started upgrade as it was EOD.
Guess what, the update failed and it messed up his system. Now the system boots but crashes after few min. Docker won't start.4
I read learncpp.com and in parallel i watched a lot of the tutorials by MakingGamesWithBen, which helped me a lot. This basically thought me C++ in less than half a year and since then i am gathering experiences and i never have problems with the language. All i have to do is look up more specific stuff like special containers or functions
I have done some experiments on my server in the past. It's a great way to learn new things. However, I am bound to make some mistakes and over time the sever becomes messier and messier.
A week ago I installed UNRAID on my machine and I love it! I can now have my critical infrastructure live and working in docker containers and vms.
Then if I want to do an experiment I spin up a VM in a couple of minutes to do my thing and remove it when I am done. No traces left!
Is it just me or others too feel that docker completely nailed it with their naming convention.
Like docker, containers etc are so aptly named by their functionality.1
Is there any kind of protocol/method where I can use something like docker containers in order to "host" compilers like gcc and use that with vscode to compile and assemble source code?
No I'm not talking about volumes (it's a bit tedious if I want to use it to manage numerous projects)3
So this week should be interesting. I am working on a (potentially) very large project for my current client and need to build a service that somewhat replicates the functionality of heroku (in that it needs to be able to load an app built in one of several languages, and spin it up in a docker container).
Unlike Heroku, however, each application also needs to be able to have a list of public and private (internal only) API routes listed and be able to dynamically route requests to the correct routes on in those containers. (Sorry if this is confusing)
Does this sound challenging and amazing? Absolutely! Do I think I may be in over my head? Yes, yes I do.
Has anyone ever built or worked with something similar?1
Hi everyone, I’m trying to wrap my head around dev ops but struggling with the whole continuous integration workflow. From my understanding it goes something like the following:
1. pushing a change to some repository (git)
2. Some tool (Jenkins) tries to build it and if it’s succeed, creates a image /container (docker).
These containers are hosted on some cloud service (aws)
Some workflow, walkthrough, or examples would be very appreciate.7
Why the heck does everyone thnk it's a good idea to run docker Containers inside a vm?
Or Containers in general...
This is unneeded redundancy and it kills the performance aspect, which makes containers favorable in comparison to vms...3
Career fair was today at my college. I don’t know why but I makes me kinda happy that some things I’m learning on my own were mentioned that aren’t taught in class
Many people mentioned micro services, some talked about containers like docker, and I had a ten minute conversation just about react js. All things I’m reading and toying with on my own time since they’re not in the curriculum, but I’ve heard them mentioned often enough in articles to know I should know about them
Making distributed scheduler that queue and run tasks on containers or other executors in future and also pulls new tasks from defined git repositories.
Tasks are added based on simple yaml configuration.
Need that for my side projects that gather data from multiple sources from time to time.
k8s looks to heavy for that and airflow can’t be configured like I want it to be so I started writing my own on Monday.
Nearly finished poc version.2
Does somebody have any experience with LXC/LXD containers on servers? I basicaly want all the services separated, but still have an easy way to manage the networking/routing for all the services and containers.
Any reference, guide or tool that helped you in master this subject?
Thank you in advance!5
Before my area of focus was the frontend and then my boss thought my skills were good enough to handle servers, now I'm here and working on the staging and production containers... Mother docker awesome...6
Oh there are lots of good tools to handle different screen sizes like containers, anchors, scaling and all that good stuff huh ?
Sets it to stretch mode instead
Spent ages building docker containers for all our projects. All worked great and I released it. So someone added a new dependency that required an Ubuntu package to be installed but forgot to tell me. My colleague said "docker isn't very stable is it" when it didn't run due to said dependency. It's a Linux container ffs, see if you blame Virtualbox under the same circumstances you idiot
Can you start services manually when using docker compose?
As in having container A, B and C with an executable A, B and C (originating from the same source btw) where I can start/stop A, B, C independently?
All examples seem to use it as a whole thing where all containers must be up for the service to work.3
Who could explain what a container is? Like a Docker container, but not necessarily Docker. Google doesn't provide simple yet meaningful answer (or there is something wrong with me today). Any ideas?8
Any recommendations for decent 13 inch laptop under 1400$ USD
- Atleast 8 gigs of RAM and option to upgrade
- resolution >= full HD
- storage >= 256 GB SSD
- battery >= 6 hrs heavy chrome usage
- Linux compatibility
- Cpu- nothing heavy - browser and docker containers, no building or compiling18
Note to self:
Variadic C++ templates combined with obfuscated combinations of stl containers put you on the right path to be a "compiler message decryption" archmage.
Especially when you use MSCV...
Do we still need virtualenv when we have containers? Cuz one of my friends thinks it’s a good idea to use virtualenv everywhere, even in production, even in an LXC?
Is it being paranoid or really a solution to some problem that would otherwise rise? I’m just curious.1
I was finally able to set up a server with my coworker's help in our work office. After struggling trying to get CoreOS to work, I quit and just went with Ubuntu Server 16.04. It's a SFF PC that we had lying around. The purpose is to run applications in Docker containers for QA, demo, and performance testing. I can say it was truly a productive day...
Bit of a stupid oopsie I had today that someone might appreciate.
We’re working on a microservice project in Spring Boot, running in a docker swarm. Past few days I get a Spring Cloud config server going in separate stack, create an overlay network, and get CI deployments to use the right profiles etc. It’s looking great, and the first component is working spectacularly.
Now just to do the other 6. Move config files to the Git repo, tweak CI, all the other faffing and hoohas; and deploy. Health checks keep failing, the containers are murdering themselves and resurrecting ad infinitum. They’re doing this so quickly that by the time I get the container ID to exec in and curl health, it’s no longer running. Cue frustration, increased caffeine and nicotine consumption; my sanity is slipping.
No errors in the logs, because from experience the Cloud Config errors ar at debug level. Whhhyyyy?? Some time later (way longer than it should have been) I realize I had never actually included the Spring Cloud Config starter. Boot 101, get your starter!
Since config client is just additional setup in properties.yml, there’s no issue of the dep isn’t there, it just doesn’t try to get the config.
The containers are still unhealthy, I can hear them screaming. But now at least it’s about something else...
Setting up dockerfile with ENV(from Visual Studio) was such a stressful endeavour from the point of view of someone that doesn't work daily with containers that I'm wondering how you master folks of net core and docket live and breath.
The gotcha was to put the ENVs at the first FROM from which the running environment was going to execute the WebAPI and not later on where there is the ENTRYPOINT point
If someone can shed some light on this behavior, would be appreciated:
I am running a couple of docker containers with lighttpd on my server (lighttpd is also installed on the host server for reverse proxy). Now whenever I kill lighttpd on my host server it also kills ALL the running lighttpd instances in my docker containers. Isn't docker supposed to be, idk, CONTAINERIZED?2