Operations | Monitoring | ITSM | DevOps | Cloud

Don't miss the blind spots: API monitoring for digital resilience

In today's digital world, applications are the lifeline of businesses. They're the engines powering everything from e-commerce transactions (think adding items to your shopping cart) to internal communication tools (imagine sending a message to a colleague). Any glitch or outage in these applications can have a domino effect, impacting revenue, productivity, and even brand reputation.

Provide full context to reviewers by including pipeline artifacts within the pull request

The code insights functionality in Bitbucket Cloud provides a variety of reports, annotations, and metrics to help your team have full context during the code review process. With code insights, you can automatically have artifacts such as static analysis reports, security scan results, artifact links, unit test results, and build status updates appear in your pull request screen so reviewers have access to all reports and statuses before they approve the code change.

What is a DevOps engineer? A look inside the role

DevOps engineers play a vital role in modern software organizations, helping to bridge the gap between software development and IT operations. DevOps is a cultural and technical approach that emphasizes collaboration, automation, and continuous integration and delivery (CI/CD). DevOps engineers are responsible for implementing and supporting these practices to improve efficiency, enhance software quality, and accelerate delivery times.

Guide to Monitoring Your Apache Zipkin Environment Using Telegraf

Using Apache Zipkin is important because it provides detailed, end-to-end tracing of requests across distributed systems, helping to identify latency issues and performance bottlenecks. Monitoring your Zipkin environment is crucial to ensure the reliability and performance of your tracing system, allowing you to quickly detect and address any anomalies or downtime.

Platform Engineering Best Practices: Data Security and Privacy

Security is and will always be a huge concern, and Platform Engineering is here to stay: so, what are some Platform Engineering best practices that can support your data security and privacy efforts? You’d be surprised where they overlap, and what you can learn about putting security and productivity together — we’ll explain.

Bridging the gap: Integrating network and application monitoring for complete visibility

As technology progresses and applications become more intertwined, sticking to the old ways of monitoring networks separately just doesn’t cut it anymore. Network and application teams often work in silos, using different tools and focusing on different goals. This split approach frequently leaves both sides with a piecemeal understanding of issues, making it challenging to pinpoint and fix problems that span both areas.

Independent, Involved, Informed, and Informative: The Characteristics of a CoPE

As our Field CTO Liz Fong-Jones says, production excellence is important for cloud-native software organizations because it ensures a safe, reliable, and sustainable system for an organization’s customers and employees. A CoPE helps organizations cultivate the practices and tools necessary to achieve that consistently. In part one of our CoPE series, we analogized the CoPE with safety departments.

Announcing HAProxy 3.0

Here we are in our twenty-third year, and open source HAProxy is going strong. HAProxy is the world’s fastest and most widely used software load balancer, with over one billion downloads on Docker Hub. It is the G2 category leader in API management, container networking, DDoS protection, web application firewall (WAF), and load balancing.

Grafana Loki query acceleration: How we sped up queries without adding resources

As we discussed when we rolled out the latest major release of Grafana Loki, we’ve grown the log aggregation system over the past five years by balancing feature development with supporting users at scale. A big part of the latter has been making queries much faster — and that was a major focus with Loki 3.0 too. We’ve seen peak query throughput grow from 10 GB/s in our Loki 1.0 days to greater than 1 TB/s even before 3.0.