Software reliability is a big deal, especially at the enterprise level, but too often companies are flying blind when it comes to the overall quality and reliability of their applications. It seems like every week, there’s a new report in the news calling out another massive software failure. Sometimes it’s just a glitch on social media causing usability issues, and other times it’s a serious issue in an aircraft system that leads to deadly crashes.
Development and IT Ops teams commonly find themselves in a game of tug-of-war between two key objectives: driving innovation and maintaining reliable (i.e. stable) software. To drive innovation, we’re seeing the emergence of CI/CD practices and the rise of automation. To maintain reliability software, DevOps practices and Site Reliability Engineering are being adopted to ensure the stability of software in fast-paced dev cycles.
When it comes to managing containers and the cluster infrastructure they run on, what’s the right tool for you?
There comes a time in the life of any application that small things we let slide become unignorable issues. For us, it happened when, after years of writing and executing code, our DB’s free disk space started to run out. Each passing day brought us closer to our eventual doom, and we finally had to allocate the time to fix the problem.
“The microservices trend is becoming impossible to ignore,” I wrote in 2016. It’s still true, although it’s certainly grown to more than just a passing fad. Back then, many would have argued this was just another unbearable buzzword, but today many organizations are reaping the very real benefits of breaking down old monolithic applications, as well as seeing the very real challenges microservices can introduce.
Demands of the competitive global economy have placed a strong emphasis on quality across the IT industry, and it shows no signs of going away. Meeting the customer’s expectations at a high degree of conformance is no longer expected to come at a premium – it is just expected.