Do all the things like ++ or -- rants, post your own rants, comment on others' rants and build your customized dev avatarSign Up
Get a devDuck
Rubber duck debugging has never been so cute! Get your favorite coding language devDuckBuy Now
Search - "uptime"
First off, a Merry Christmas to everyone who celebrates, happy holidays to everyone, and happy almost-new-year!
Tim and I are very happy with the year devRant has had, and thinking back, there are a lot of 2017 highlights to recap. Here are just a few of the ones that come to mind (this list is not exhaustive and I'm definitley forgetting stuff!):
- We introduced the devRant supporter program (devRant++)! (https://devrant.com/rants/638594/...). Thank you so much to everyone who has embraced devRant++! This program has helped us significantly and it's made it possible for us to mantain our current infrustructure and not have to cut down on servers/sacrifice app performance and stability.
- We added avatar pets (https://devrant.com/rants/455860/...)
- We finally got the domain devrant.com thanks to @wiardvanrij (https://devrant.com/rants/938509/...)
- The first international devRant meetup (Dutch) with organized by @linuxxx and was a huge success (https://devrant.com/rants/937319/... + https://devrant.com/rants/935713/...)
- We reached 50,000 downloads on Android (https://devrant.com/rants/728421/...)
- We introduced notif tabs (https://devrant.com/rants/1037456/...), which make it easy to filter your in-app notifications by type
- @AlexDeLarge became the first devRant user to hit 50,000++ (https://devrant.com/rants/885432/...), and @linuxxx became the first to hit 75,000++
- We made an April Fools joke that got a lot of people mad at us and hopefully got some laughs too (https://devrant.com/rants/506740/...)
- We launched devDucks!! (https://devducks.com)
- We got rid of the drawer menu in our mobile apps and switched to a tab layout
- We added the ability to subscribe to any user's rants (https://devrant.com/rants/538170/...)
- Introduced the post type selector (https://devrant.com/rants/850978/...) (which will be used for filtering - more details below)
- Started a bug/feature tracker GitHub repo (https://github.com/devRant/devRant)
- We did our first ever live stream (https://youtube.com/watch/...)
- Added an awesome all-black theme (devRant++) (https://devrant.com/rants/850978/...)
- We created an "active discussions" screen within the app so you can easily find rants with booming discussions!
- Thanks to the suggestion of many community members, we added "scroll to bottom" functionality to rants with long comment threads to make those rants more usable
- We improved our app stability and set our personal record for uptime, and we also cut request times in half with some database cluster upgrades
- Awesome new community projects: https://devrant.com/projects (more will be added to the list soon, sorry for the delay!)
- A new landing page for web (https://devrant.com), that was the first phase of our web overhaul coming soon (see below)
Even after all of this stuff, Tim and I both know there is a ton of work to do going forward and we want to continue to make devRant as good as it can be. We rely on your feedback to make that happen and we encourage everyone to keep submitting and discussing ideas in the bug/feature tracker (https://github.com/devRant/devRant).
We only have a little bit of the roadmap right now, but here's some things 2018 will bring:
- A brand new devRant web app: we've heard the feedback loud and clear. This is our top priority right now, and we're happy to say the completely redesigned/overhauled devRant web experience is almost done and will be released in early 2018. We think everyone will really like it.
- Functionality to filter rants by type: this feature was always planned since we introduced notif types, and it will soon be implemented. The notif type filter will allow you to select the types of rants you want to see for any of the sorting methods.
- App stability and usability: we want to dedicate a little time to making sure we don't forget to fix some long-standing bugs with our iOS/Android apps. This includes UI issues, push notification problems on Android, any many other small but annoying problems. We know the stability and usability of devRant is very important to the community, so it's important for us to give it the attention it deserves.
- Improved profiles/avatars: we can't reveal a ton here yet, but we've got some pretty cool ideas that we think everyone will enjoy.
- Private messaging: we think a PM system can add a lot to the app and make it much more intuitive to reach out to people privately. However, Tim and I believe in only launching carefully developed features, so rest assured that a lot of thought will be going into the system to maximize privacy, provide settings that make it easy to turn off, and provide security features that make it very difficult for abuse to take place. We're also open to any ideas here, so just let us know what you might be thinking.
There will be many more additions, but those are just a few we have in mind right now.
We've had a great year, and we really can't thank every member of the devRant community enough. We've always gotten amazingly positive feedback from the community, and we really do appreciate it. One of the most awesome things is when some compliments the kindness of the devRant community itself, which we hear a lot. It really is such a welcoming community and we love seeing devs of all kind and geographic locations welcomed with open arms.
2018 will be an important year for devRant as we continue to grow and we will need to continue the momentum. We think the ideas we have right now and the ones that will come from community feedback going forward will allow us to make this a big year and continue to improve the devRant community.
Thanks everyone, and thanks for your amazing contributions to the devRant community!
Looking forward to 2018,
- David and Tim52
User:"It's not working"
Me:"Have you turned it off and on again?"
Me: goes down there, system uptime is 360days...
"How do you turn it off?"
User:"by pressing the button on the monitor17
Merry Christmas to everyone who celebrates, happy holidays to everyone, and happy almost-new-year!
We had a bit of a slow year in terms of devRant updates, but we gained some momentum towards the end of the year and we're looking forward to carrying it into 2020. Recently, we launched what I think are our coolest new avatar items yet (https://devrant.com/rants/2322869/...) and behind the scenes we got our iOS/Android apps on the latest version of the frameworks we use, which will help us continue to improve stability. Still, we definitely would have liked to do more, but we're optimistic the coming year will bring great things for devRant.
One thing we are very proud of is this year we had our best year ever in terms of platform stability and uptime. Despite the platform growing and our userbase growing, we had almost no complete app downtime even though our infrastructure is minimal. A large part of this is thanks to devRant++ supporters, who allow us to maintain a small but effective tier of infrastructure and redundancy.
In the coming year, we're going to launch one of our most ambitious initiatives yet, and we're also going to continue to improve the devRant experience itself. We want to try to gather more user feedback, so we'll be working on a way to do that too. Stay tuned, more on this stuff coming soon.
As always, thank you everyone, and thanks for your amazing contributions to the devRant community! And thank you to our awesome devRant++ supporters for continuing to be the main drivers to keeping devRant up and running.
Looking forward to 2020,
- David and Tim31
2 years ago: Connection goes down at the office
Boss: -"Damnit, you are responsible to ensure uptime. Fix a redundant connection asap."
Me: Fixes redundant connection
Today: Connection goes down at the office, failover connection does not work.
Me: Calls ISP and asks what's happened.
ISP: -"Your boss cancelled the account 3 months ago"
Mainly a story with some ranting sprinkled on top :)
My grandma used to refuse to turn her PC off because ONE SINGLE WINDOWS 8.1 UPDATE like 2 years ago made it so that Windows would turn the PC on by itself, nevermind it was fixed in the next update, she would simply refuse to try to turn the PC off from that point on... I never looked at the uptime but the PC was sitting on a carpet and all the fans were struggling to keep spinning, so much dust! When I asked her why it's so dusty all around the PC (the table that the PC was under was covered in a really thick layer of dust) she replied that the chambermaid comes every Wednesday to clean up the whole flat so she has no idea. I set up a camera that was recording the whole room so I could see what was going on... sure enough the chambermaid avoided anything that looked like it could have some ICs in it with a distance of at least 2 meters... When I showed the video to my grandma she was like: Yes, she is scared of electronics...
Seriously fuck me... Allright, I cleaned the whole table, carpet and the inside of PC, everything ran so quietly that I wasn't sure if I plugged the fans in.
Anyway after this I had to record the PC one whole night so I could prove to my grandma that it would not turn on by itself and she can in fact turn it off.
Fuck me... what a bad week that was...
PS: Yes, my stories do not have a point :)5
"Arch Linux is actually not that difficult".
I ssh'ed into my home server yesterday.
I was greeted by a message from an ext3 disk about needing fsck. Fine, "I haven't been in here for a while, might as well do some maintenance". fsck /dev/sda6, let's go!
This nicely "repaired" the sshd service (i.e. cleared the sectors), I cursed at myself for pressing enter at "repair (y)" right before the connection broke.
So I connected a display and keyboard... ok so let's just pacman -Sy sshd or whatever. We can do this! Just check the wiki, shouldn't be that hard!
Wait... pacman has not run since 2010? WAIT IT'S ACTUAL UPTIME IS 9 YEARS??? I guess we know why I'm a DB admin and not devops...
Hmm all the mirrors give timeouts? Oh. The i686 processor architecture isn't even supported anymore...?
4 hours, 11 glasses of cognac, 73 Arch32 wiki/forum pages, 2 attempts at compiling glibc, and 4 kernel panics later: "I think I'll buy a new server".18
Last half hour of last Friday my Cinnamon desktop started fucking up every time I booted after like 20 seconds of uptime.
Logs said something about xorg errors but after like 15 reboots I said fuck it.
Was quite nervous this morning because of that issue and possibly not being able to open a terminal for installing a different DE before everything started to get screwed again.
Booted up. No freezes. No errors. It just works.
I guess my computer needed weekend too 😋14
We've got a team of around 20 developers and the most junior of them all is a interesting specimen.
The kind of person who thinks they a 'expert' in anything and everything and is constantly trying to school our senior developers who have 20+ years experience behind them.
The sort of person that spends 15 seconds googling something he has never heard of before, but now that he has skimmed 1 page on Google would classify himself as a 'expert' in said topic.
He comes into my office yesterday and proclaims that it has been decided by himself that he no longer wants to be a developer anymore and wants to do Ops/Infrastructure, then starts rambling on about how he is a Kubernetes expert.
I asked what experience he had with Kubernetes and his response was "I watched a webinar they did last night" to which I asked if he had ever actually used anything to do with Kubernetes in his life.
"No, but I'll watch a few YouTube videos and will then be more than qualified" he says
Followed by him telling me that we'll be moving all of our current Docker Swarm clusters into Kubernetes.
This was news to me (I'm head of infrastructure and operations)
I needed a good giggle, so I asked why we would get rid of our exisiting Docker infrastructure that's got a 100% uptime over the past 2 years and has worked without failure. It's truely been a dream.
He says "Because it's shiny and cool and better"
The nest afternoon he comes to me and says "When I move everything into Kubernetes I am going to convert everything into micro services"
He says that he watched a YouTube video the night before on microservices and has decided that it's what we need to use for a particular project.
(It's a simple php website that gets 100 hits per day)
Hopefully his boss will notice that he is producing no output soon. Don't want to tell the manager that the guy he hired delivers no work and lives in a fantasy land.
"your not touching the infrastructure. Ever"16
Well, the OS that is my brain has an official uptime of 19 years... It's gotten pretty buggy, I'm not sure what the devs were thinking with half the features...13
So I work in the service desk and the moment and one of our clients use Mac's.
One of the end users called up saying that it was being slow and sluggish.
End user: hi my Mac is being slow.
Me: when did you last reboot it?
End user: last night
*Runs uptime in terminal*
Me: are you sure you rebooted it last night?
End user: yes I close the lid every night...
The up time was 68 days...9
I've found and fixed any kind of "bad bug" I can think of over my career from allowing negative financial transfers to weird platform specific behaviour, here are a few of the more interesting ones that come to mind...
#1 - Most expensive lesson learned
Almost 10 years ago (while learning to code) I wrote a loyalty card system that ended up going national. Fast forward 2 years and by some miracle the system still worked and had services running on 500+ POS servers in large retail stores uploading thousands of transactions each second - due to this increased traffic to stay ahead of any trouble we decided to add a loadbalancer to our backend.
This was simply a matter of re-assigning the IP and would cause 10-15 minutes of downtime (for the first time ever), we made the switch and everything seemed perfect. Too perfect...
After 10 minutes every phone in the office started going beserk - calls where coming in about store servers irreparably crashing all over the country taking all the tills offline and forcing them to close doors midday. It was bad and we couldn't conceive how it could possibly be us or our software to blame.
Turns out we made the local service write any web service errors to a log file upon failure for debugging purposes before retrying - a perfectly sensible thing to do if I hadn't forgotten to check the size of or clear the log file. In about 15 minutes of downtime each stores error log proceeded to grow and consume every available byte of HD space before crashing windows.
#2 - Hardest to find
This was a true "Nessie" bug.. We had a single codebase powering a few hundred sites. Every now and then at some point the web server would spontaneously die and vommit a bunch of sql statements and sensitive data back to the user causing huge concern but I could never remotely replicate the behaviour - until 4 years later it happened to one of our support staff and I could pull out their network & session info.
Turns out years back when the server was first setup each domain was added as an individual "Site" on IIS but shared the same root directory and hence the same session path. It would have remained unnoticed if we had not grown but as our traffic increased ever so often 2 users of different sites would end up sharing a session id causing the server to promptly implode on itself.
#3 - Most elegant fix
Same bastard IIS server as #2. Codebase was the most unsecure unstable travesty I've ever worked with - sql injection vuns in EVERY URL, sql statements stored in COOKIES... this thing was irreparably fucked up but had to stay online until it could be replaced. Basically every other day it got hit by bots ended up sending bluepill spam or mining shitcoin and I would simply delete the instance and recreate it in a semi un-compromised state which was an acceptable solution for the business for uptime... until we we're DDOS'ed for 5 days straight.
My hands were tied and there was no way to mitigate it except for stopping individual sites as they came under attack and starting them after it subsided... (for some reason they seemed to be targeting by domain instead of ip). After 3 days of doing this manually I was given the go ahead to use any resources necessary to make it stop and especially since it was IIS6 I had no fucking clue where to start.
So I stuck to what I knew and deployed a $5 vm running an Nginx reverse proxy with heavy caching and rate limiting linked to a custom fail2ban plugin in in front of the insecure server. The attacks died instantly, the server sped up 10x and was never compromised by bots again (presumably since they got back a linux user agent). To this day I marvel at this miracle $5 fix.1
My phone is my most reliable computer. It made it to 2000 hours of uptime before I restarted it to update.
Props to Android.4
Oh Christ.. just been looking for hosting companies here in Belgium to look for sysadmin positions.. one of the fucking companies posted this: "we provide Uptime-as-a-Service"
The fucking cringe!!! Uptime as a service! Everything including the only fucking job a hoster has, keeping shit up and running.. as a service.. fuck!4
TL; DR: Bringing up quantum computing is going to be the next catchall for everything and I'm already fucking sick of it.
Actual convo i had:
"You should really secure your AWS instance."
"Isnt my SSH key alone a good enough barrier?"
"There are hundreds of thousands of incidents where people either get hacked or commit it to github."
"Well i wont"
"Just start using IP/CIDR based filtering, or i will take your instance down."
"But SSH keys are going to be useless in a couple years due to QUANTUM FUCKING COMPUTING, so why wouldnt IP spoofing get even better?"
"Listen motherfucker, i may actually kill you, because today i dont have time for this. The whole point of IP-based security is that you cant look on Shodan for machines with open SSH ports. You want to talk about quantum computing??!! Lets fucking roll motherfucker. I dont think it will be in the next thousand years that we will even come close to fault-tolerant quantum computing.
And even if it did, there have been vulnerabilities in SSH before. How often do you update your instance? I can see the uptime is 395 days, so probably not fucking often! I bet you "dont have anything important anyways" on there! No stored passwords, no stored keys, no nothing, right (she absolutely did)? If you actually think I'm going to back down on this when i sit in the same room as the dude with the root keys to our account, you can kindly take your keyboard and shove it up your ass.
Christ, I bet that the reason you like quantum computing so much is because then you'll be able to get your deepfakes of miley cyrus easier you perv."10
Logged into a legacy box with an uptime longer than my age. That's how the troubleshooting went the rest of the day.
Question again about hosting web apps, has anyone using the paid bills for heroku? Since they gave 1000 hours uptime for their free version, is it possible to still use it for a web apps? And mix the bills for another web apps? Or is it better to host under vps?23
Welp, its official, with Debian Buster adoption into our mainline, we are officially switching from Sys-V-Init to SystemD.
I still do not know how I feel about it.
From the professional point of view - Its a relief. SystemD has so many more neat features that make the life of a sysadmin easier. If any, I love that it tracks the uptime of a service, making it incredibly easy the last time it crashed / restarted...
On the other... I just... Am kind of afraid where the whole systemd environment will go with time... And... I guess... I am also worried about how much systemd is taking over in the system itself... It will mean learning quite a few new services, debugging routines and such...
A new era of GNU/SystemD/Linux is upon us.16
Just checked my laptop uptime and it's been on one charge for exactly 1 month, with average use and aggressive tlp/powertop settings 😰18
100% uptime claim on this Minecraft server host.
Me, a server software support helper: no, that guarantee is a lie. Check their SLA.
User: nah they don't lie, it's ALWAYS up.
Ugh, people. 100% uptime means 0 minutes downtime over a year. You're kidding, right? Not even governments strive for that level of uptime. I think 6 nines is the highest I've seen.14
All the Linux servers I manage:
Uptime 300+ days
All the Linux server I manage inside hyper-v managed by our it:
Uptime max 7 days...
Wtf? Do you really have to restart the host machine once a week?8
How a linux sysadmin has sex :
who | grep -i hot && grep -i female | date; cd ~; unzip; touch; strip; finger; mount; gasp; yes; uptime; umount; sleep;
Unnamed hacking game - "terminal" graphics
-Multiplayer. Last man standing.
-Like a tower-defence game but technical
You work for a company that has outsourced their technical department to Bykazistan, a country with good internet and bad laws. On one hand, labor is very cheap! There are no pesky laws protecting workers, so you don't need to pay them what they're worth. Phew. However, there are also no laws against cyber crime. But for a start-up like you, the risk is worth the reward!
...which would be great! If you were the only company with that idea. As it turns out, you aren't. All of your competitors also recently outsourced to Bykazistan, and that could be an issue.
You would be afraid, but you are a hardened businessman. You are familiar with the cut-throat nature of the business world and where others see risk, you see opportunity. Let the games begin.
Your mission is to protect your ciritical assets at all costs, eliminate your opponents, and make ciritical financial decisions - all while maintaining your uptime!
Build a botnet and attack your competition to decrease their uptime and disable their attacks. Port scan your opponents to learn more about their network, but beware of honeypots! Initiate devastating social engineering attacks - and train your employees against them! Brute-force their credentials, and strengthen your own.
Make sure to keep your software patched...6
A while ago I stumbled upon this cheap vps provider called dedistation (lowest was $15/year), so I end up buying a one year subscription and transferred all of my personal sites to it. Fast forward six month later, uptime robot notified me that my vps is down. No problem, I'll just submit a ticket. Few days go by and I yet to receive a response. Not a problem, I'll just try and login via the serial interface and get my shit, no luck there either.
Seems like these motherfuckers just packed up and went offline without a shit given.
No response, no notification untill today! How more twat a company could be?
Lessons learned the hard way
- always backup regularly
- check and transfer nameserver or no emails for days.
- you get what u pay for (haven't learned this fully yet. Went again with a cheap legit-ish provider (arubacloud)7
Closed the lid on my laptop. It goes to standby and the sleep light comes on. Unplug from wall. Sleep light turns off. Open it up nothing. Push power button, boots up like nothing happens. What the damn hell.
I really blow my lid off when shit like this happens. Like toddler temper tantrum level. Lost my uptime, probably lost some work, need to open everything up again, and in ten years of owning this computer this has happened once. Which begs the question... just why. Did I remove the ac power at exactly the wrong moment when it was switching over or something?1
Debian + i3wm running on the WSL just fine :D
OS: Debian GNU/Linux 9.5 (stretch) on Windows 10 x86_64
Uptime: 10 seconds
Shell: bash 4.4.12
CPU: Intel i5-3570 (4) @ 3.4GHz
Memory: 4025MB / 8158MB7
Okay, I have a lot to rant about discord, but today, exceptionally, to the point.
I have my dedicated server. It has uptime last 3 years better than 99,99% (was down 15ish minutes for maintainance and RAM upgrade and like 10 minutes down becouse hoster's generators failed to trigger when there was outage)
This year it was up 24/7/365.
Why am I saying it?
Well, my TS3 server is up 100% of time this year. Yet still everyone moves to discord and suffers brutal audio quality and audio lags, and outages like right now. Its not first time this year and recently discord was acting up before. Today they scored bigger downtime than my dedi server (thats not redundant, not distributed nor any fancy "uptime helpers") last 3 years.
Why the fuck people prefer discord to ts3 other that it allows to upload images more conviniently? Okay, it looks nicer, and is like 10 times heavier on machine, but other than that? Its beyond me.
E: fix typo
E2: fix typo29
Walking around in naturvetenskapliga museumet (a museum in Stockholm, Sweden) and i see this (image below).
This information displays which seems to run Windows restarted. I would say that it happend because of the long uptime these machines have and Windows sucks with having long uptime.
Solution: Run it on a Linux based operating system instead.
I been using digital ocean to host my server for a project, but they seem to get shutdown because of DoS behaviour. I have no idea why. The server is doing some soap and rest communication and controlling a database.
To be fair the password was poor, but it was meant to be a fast way for four people to work on it at the same time.
But after the first shutdown, we rebuild the server and work on functions. Finish the work and went home. But in the server 9 hours of uptime with 2 of them unsupervised it was detected as DoS behaving server.
I'm actually setting up a server that's supposed to have a very good uptime with Arch Linux
I use Arch btw13
just installed Pop! Os moving from Ubuntu 18.04 on my desktop. damn Ubuntu is so bloated, a lot of halt (freeze) spiking, cannot recommend it to anyone anymore.
Currently installing Proxmox on my server moving from Ubuntu 16.04 Server, recorded 4126 Hour of uptime. Damn it's gonna be a loooooong day3
That day when you need to restart your vsphere virtualizer after 2 and 1/2 years of uptime, just to find it won't boot in any possible way, and you rush to install it fresh on an USB drive 20 mins before your sla runs out... that day was today, fml!
who | grep -i blonde | date; cd ~; unzip; touch; strip; finger; mount; gasp; yes; uptime; umount; sleep
Met a client. He need some abc services. Client asked me what will be the sla uptime. I told him unintentionally that it will be 99% instead of 99.99% and he was like that's good. I mean seriously....?
Was he okay with the 3 days 15 hrs downtime?5
Amazon's stock price is a better uptime indicator than their fucking status page.
Ok so there has been a lot of outages this month with AWS going down justeat hungry house some big ones and people have lost their minds my rant is at what point did we fool normal people into thinking the web has 100% uptime and never has the possibility of going down for 10minutes.1
Just checked the uptime of my pc, thougt about 5-6 days of uptime... wtf happend to me the last 3 weeks?2
A side project lingering around is building a .NET Core based GUI program to monitor uptime and health of various Windows and Linux servers. I'm aware there are other projects that could do the same thing but I'm wanting to do this as a lesson in C# and cross-platform coding (I plan this to work on both Windows and Linux).
The program is currently CLI based on Windows with functionality to configure it and it's behaviour via config file, it currently sends email via SMTP to a specified email recipient to notify if there has been outages or performance degradation.
But of course University is in the way as well as work. Oh well... maybe I'll get to it in a couple months.
Why can't windows just use 'uptime' to show how long the site has been up rather than having to fucking waiting for 'systeminfo' to load.... I'm considering switching over to Linux in the new year, or just buy a new laptop and put Linux on it and use it as my daily driver.4
Branch Staff Member: EFTPOS isn't working
Me: Are you logged in as the right user?
*staff switches user*
Staff: Still not working
*remote in and see software had crashed*
*check computer uptime* ... 38 days
Me: Give it a restart
Rant: why do people expect their issues to be resolved magically by IT when rebooting would (usually) fix it?