Operations | Monitoring | ITSM | DevOps | Cloud

August 2023

Data Lake vs Data Warehouse

Data warehouses and data lakes represent two of the leading solutions for enterprise data management in 2023. While data warehouses and data lakes may share some overlapping features and use cases, there are fundamental differences in the data management philosophies, design characteristics, and ideal use conditions for each of these technologies.

Dashboard Studio: How to Configure Show/Hide and Token Eval in Dashboard Studio

You may be familiar with manipulating tokens via `eval` or `condition`, or showing and hiding panels via `depends` in Classic (SimpleXML) dashboards, and wondering how to do that in Dashboard Studio. In this blog post, we'll break down how to accomplish these use cases in Dashboard Studio, using the same examples that were shown at.conf23.

What's New with Fluentd & Fluent Bit

At the recent KubeCon EU, we learned the significant news of the FluentBit v2.0 major release with numerous new features. What’s new and what’s to come for this key log aggregation tool? On the latest OpenObservability Talks, I hosted Eduardo Silva, one of the maintainers of Fluentd, a creator of Fluent Bit and co-founder of Calyptia.

Log Data 101: What It Is & Why It Matters

If you mention 'log data' at a crowded business event, you'll quickly be able to tell who's in IT and who isn't. For the average person, log data is about as thrilling as a dental appointment or reconciling a years-old bank account. At the mere mention of log data, their eyes glaze over as they search for an escape from the conversation. Conversely, IT professionals' eyes light up and they become animated when the topic of log data arises.

Operational Intelligence: 6 Steps To Get Started

The ability to make decisions quickly can mean the difference between success and stagnation. Of course, quick decisions aren’t necessarily the right decisions. The right decisions are the best informed, and the best way to get informed is through data. That’s what operational intelligence is all about. In this article, we’re diving into all things operational intelligence (OI), including key benefits, goals and how to get started.

Incident Management Today: Benefits, 6-Step Process & Best Practices

Disruptive cybersecurity incidents become more and more commonplace each day. Even if nothing is directly hacked, these incidents can harm your systems and networks. Navigating cybersecurity incidents is a constant challenge — the best way to stay ahead of the game is with effective incident management.
Sponsored Post

Serverless Elasticsearch: Is ELK or OpenSearch Serverless Architecture Effective?

Here's the question of the hour. Can you use serverless Elasticsearch or OpenSearch effectively at scale, while keeping your budget in check? The biggest historical pain points around Elasticsearch and OpenSearch are their management complexity and costs. Despite announcements from both Elasticsearch and OpenSearch around serverless capabilities, these challenges remain. Both of these tools are not truly serverless, let alone stateless, hiding their underlying complexity and passing along higher management costs to the customer.

Distributed Systems Explained

Distributed systems might be complicated…luckily, the concept is easy to understand! A distributed system is simply any environment where multiple computers or devices are working on a variety of tasks and components, all spread across a network. Components within distributed systems split up the work, coordinating efforts to complete a given job more efficiently than if only a single device ran it.

Why Observability Architecture Matters in Modern IT Spaces

Observability architecture and design is becoming more important than ever among all types of IT teams. That’s because core elements in observability architecture are pivotal in ensuring complex software systems’ smooth functioning, reliability and resilience. And observability design can help you achieve operational excellence and deliver exceptional user experiences. In this article, we’ll delve into the vital role of observability design and architecture in IT environments.

The Quirky World of Anomaly Detection

Hey there, data detectives and server sleuths! Ever find yourself staring at a screen full of numbers and graphs, only to have one data point wave at you like a tourist lost in Times Square? Yup, you’ve stumbled upon the cheeky world of Anomaly Detection—where data points act more mysterious than your cat when it suddenly decides to sprint around the house at 2 AM. So buckle up!

Coralogix Deep Dive - Remote Query for Logs

In this video, we'll explore the functionality and best practices with Coralogix Remote Query for Logs. Coralogix supports direct, unindexed queries to the archive for both logs and traces, and stores data in cloud object storage directly in the customers cloud account, making for rock bottom retention costs, blazing fast performance and outstanding scalability.

The Leading APM Use Cases

The majority of users continually depend on a variety of web applications to meet their everyday needs, so a business’s success is now often proportionate to the success of its application performance. As a result, the importance of using an appropriate APM solution has become even greater to businesses globally. Application Performance Monitoring (APM) still continues to grow in popularity and is now considered a must for observing the health and performance of your organization's applications.

Parquet File Format: The Complete Guide

How you choose to store and process your system data can have significant implications on the cost and performance of your system. These implications are magnified when your system has data-intensive operations such as machine learning, AI, or microservices. And that’s why it’s crucial to find the right data format. For example, Parquet file format can help you save storage space and costs, without compromising on performance.

Deleting Fields with BindPlane OP

Are you ingesting unnecessary fields? See how to use the "Delete Fields" processor to remove fields from your log stream in BindPlane OP. Then use our live preview capabilities to see the changes prior to rolling out to your agents. About ObservIQ: observIQ brings clarity and control to our customer's existing observability chaos. How? Through an observability pipeline: a fast, powerful and intuitive orchestration engine built for the modern observability team. Our product is designed to help teams significantly reduce cost, simplify collection, and standardize their observability data.

How to deploy Hello World Elastic Observability on Google Cloud Run

Elastic Cloud Observability is the premiere tool to provide visibility into your running web apps. Google Cloud Run is the serverless platform of choice to run your web apps that need to scale up massively and scale down to zero. Elastic Observability combined with Google Cloud Run is the perfect solution for developers to deploy web apps that are auto-scaled with fully observable operations, in a way that’s straightforward to implement and manage.

Blackhat 2023 Recap: How Will Advanced AI Impact Cybersecurity?

Ed Bailey and Jackie McGuire from Cribl will recap Black Hat 2023, focusing on emerging trends in cybersecurity, including the rise of advanced AI. We’ll share insights and anecdotes from our time at the event. Tune into the live stream for an engaging discussion, and come prepared with your thoughts and questions about Black Hat and the future of cybersecurity.

Getting started with Grafana Loki (Grafana Office Hours #09)

Senior Principal Solutions Engineer Ward Bekker talks about getting started with Grafana Loki: what Loki is, why you need log aggregation, and how it fits into the rest of the Grafana stack. He is joined by Developer Advocates Paul Balogh and Nicole van der Hoeven to tell you everything you need to know about Loki.

Mean Time to Repair (MTTR): Definition, Tips and Challenges

The availability and reliability of any IT service ultimately govern end-user experience and service performance, both of which have significant business impact. These two concepts — availability and reliability — are particularly relevant in the era of cloud computing, where software drives business operations, but that software is often managed and delivered as a service by third-party vendors.

When Two Worlds Collide: AI and Observability Pipelines

In today's data-driven world, ensuring the stability and efficiency of software applications is not just a need but a requirement. Enter observability. But as with any evolving technology, there's always room for growth. That growth, as it stands today, is the convergence of artificial intelligence (AI) with observability pipelines. In this blog, we'll explore the idea behind this merge and its potential.

What is a Real-Time Data Lake?

A data lake is a centralized data repository where structured, semi-structured, and unstructured data from a variety of sources can be stored in their raw format. Data lakes help eliminate data silos by acting as a single landing zone for data from multiple sources. But what’s the difference between a traditional data lake and a real-time data lake?

Big Data Analytics: Challenges, Benefits and Best Tools to Use

Imagine yourself with a folder containing millions of gigabytes of data. If you were asked to process it with an Excel spreadsheet, you wouldn’t need to be a data expert to know that’s impossible. We refer to that amount of data as “big data”. Big data requires advanced techniques, tools, and methods beyond what regular data analytics entails, which is where big data analytics comes in.

Coralogix vs Splunk: Support, Pricing and More

Splunk has become one of several players in the observability industry, offering a set of features and a specific focus on legacy and security use cases. That being said, how does Splunk compare to Coralogix as a complete full-stack observability solution? Let’s dive into the key differences between Coralogix vs Splunk, including customer support, pricing, cost optimization, and more.

Comparing Six Top Observability Software Platforms

When it comes to observability, your organization will have no shortage of options for tools and platforms. Between open source software and proprietary vendors, you should be able to find the right tools to fit your use case, budget and IT infrastructure. Observability should be cost-efficient, easy to implement and customers should be provided with the best support possible.

The Future of Observability: Navigating Challenges and Harnessing Opportunities

Observability solutions can easily and rapidly get complex — in terms of maintenance, time and budgetary constraints. But observability doesn’t have to be hard or expensive with the right solutions in place. The future of your observability can be a bright one.

Splunk and the Four Golden Signals

Last October, Splunk Observability Evangelist Jeremy Hicks wrote a great piece here about the Four Golden Signals of monitoring. Jeremy’s blog comes from the perspective of monitoring distributed cloud services with Splunk Observability Cloud, but the concepts of Four Golden Signals apply just as readily to monitoring traditional on-premises services and IT infrastructure.

How we scaled Grafana Cloud Logs' memcached cluster to 50TB and improved reliability

Grafana Loki is an open source logs database built on object storage services in the cloud. These services are an essential component in enabling Loki to scale to tremendous levels. However, like all SaaS products, object storage services have their limits — and we started to crash into those limits in Grafana Cloud Logs, our SaaS offering of Grafana Loki.

How to get actionable insights from your data

“When you peel back business issues, more times than not, you will find that the root cause is directly tied to data problems,” says Matthew Minetola, CIO at Elastic®. In today's world, all companies, new and old, are awash in data from multiple sources — stored in multiple systems, versions, and formats — and it’s getting worse all the time.

How to Log HTTP Headers With HAProxy for Debugging Purposes

HAProxy offers a powerful logging system that allows users to capture information about HTTP transactions. As part of that, logging headers provides insight into what's happening behind the scenes, such as when a web application firewall (WAF) blocks a request because of a problematic header. It can also be helpful when an application has trouble parsing a header, allowing you to review the logged headers to ensure that they conform to a specific format.

Code, Coffee, and Unity: How a Unified Approach to Observability and Security Empowers ITOps and Engineering Teams

In today's fast-paced and ever-changing digital landscape, maintaining digital resilience has become a critical aspect of business success. It is no longer just a technical challenge but a crucial business imperative. But when observability teams work in their own silos with tools, processes, and policies, disconnected from the security teams, it becomes more challenging for companies to achieve digital resilience.

Data Quality Metrics: 5 Tips to Optimize Yours

Amid a big data boom, more and more information is being generated from various sources at staggering rates. But without the proper metrics for your data, businesses with large quantities of information may find it challenging to effectively and grow in competitive markets. For example, high-quality data lets you make informed decisions that are based on derived insights, enhance customer experiences, and drive sustainable growth.

IT Orchestration vs. Automation: What's the Difference?

As modern IT systems grow more elaborate, encompassing hardware and software across hybrid environments, the prospect of managing these systems often grows beyond the capacity an IT team can handle. Automation is one great way to help. But it's important to know that not all automation is the same — chatbots are probably not the solution your team is looking for to handle these incredibly complex systems.

Simplifying Data Lake Management with an Observability Pipeline

Data Lakes can be difficult and costly to manage. They require skilled engineers to manage the infrastructure, keep data flowing, eliminate redundancy, and secure the data. We accept the difficulties because our data lakes house valuable information like logs, metrics, traces, etc. To add insult to injury, the data lake can be a black hole, where your data goes in but never comes out. If you are thinking there has to be a better way, we agree!

Lookup Tables and Log Analysis: Extracting Insight from Logs

Extracting insights from log and security data can be a slow and resource-intensive endeavor, which is unfavorable for our data-driven world. Fortunately, lookup tables can help accelerate the interpretation of log data, enabling analysts to swiftly make sense of logs and transform them into actionable intelligence. This article will examine lookup tables and their relationship with log analysis.

Top Elasticsearch Metrics to Monitor | Troubleshooting Common error in Elasticsearch

Monitoring Elasticsearch metrics is absolutely essential! Monitoring gives you information about the functionality, overall condition, and performance of your Elasticsearch cluster. Without monitoring, you risk missing important “red-flags” that could make your cluster inaccessible or crash. Which could result in data loss and downtime; both of which would be expensive for your company.

A Complete Guide to Tracking CDN Logs

The Content Delivery Network (CDN) market is projected to grow from 17.70 billion USD to 81.86 billion USD by 2026, according to a recent study. As more businesses adopt CDNs for their content distribution, CDN log tracking is becoming essential to achieve full-stack observability. That being said, the widespread distribution of the CDN servers can also make it challenging when you want visibility into your visitors’ behavior, optimize performance, and identify distribution issues.

Parsing logs with the OpenTelemetry Collector

This guide is for anyone who is getting started monitoring their application with OpenTelemetry, and is generating unstructured logs. As is well understood at this point, structured logs are ideal for post-hoc incident analysis and broad-range querying of your data. However, it’s not always feasible to implement highly structured logging at the code level.

What Is ITOPs? IT Operations Defined

IT operations, or ITOps, refers to the processes and services administered by an organization's IT staff to its internal or external clients. Every organization that uses computers has a way of meeting the IT needs of their employees or clients, whether or not they call it ITOps. In a typical enterprise environment, however, ITOps is a distinct group within the IT department. The IT operations team plays a critical role in accomplishing business goals.

Developing the Splunk App for Anomaly Detection

Anomaly detection is one of the most common problems that Splunk users are interested in solving via machine learning. This is highly intuitive, as one of the main reasons our Splunk customers are ingesting, indexing, and searching their systems’ logs and metrics is to find problems in their systems, either before, during, or after the problem takes place. In particular, one of the types of anomaly detection that our customers are interested in is time series anomaly detection.

Exploring & Remediating Consumption Costs with Google Billing and BindPlane OP

We’ve all been surprised by our cloud monitoring bill at one time or another. If you are a BindPlane OP customer ingesting Host Metrics into Google Cloud Monitoring, you may be wondering which metrics are impacting your bill the most. You may have metrics enabled that aren’t crucial to your business, driving unnecessary costs. How do we verify that and remediate?

BindPlane OP Architecture Overview

In this overview we dive into the BindPlane architecture focusing on the two main components. 1) BindPlane OP Server: acts as a orchestration layer that all of your agents are connected to giving you visibility into what is happening. 2) BindPlane Agent: is a distribution of the OpenTelemetry collector, sitting on your edge nodes collecting your telemetry data or acting as an aggregator (or gateway node) collecting from other edge devices and then routing to your destinations.

How Gaming Analytics and Player Interactions Enhance Mobile App Development

The number of mobile game users is expected to increase to 2.3 billion users by 2027, with a CAGR of 7.08%. The resulting projected market volume is a staggering $376.7 billion by 2027. Competition is fierce, and differentiation is key to winning out in this rapidly growing market. To understand their users and build better games, gaming companies need to use data analytics to interpret how players interact with their games. Effective use of video game data can help companies.

Introducing the Splunk App for Behavioral Profiling

Splunk is the platform for a million use cases, used to investigate operational data across security, observability, fraud, business intelligence and many other domains. But, in my time at Splunk, I’ve come to realize that all of our customers face challenges that stem from the same core problem: Within exploding data volumes, finding the anomalously behaving entities that are most threatening to the resilience of their organization.

How to Strengthen Kubernetes with Secure Observability

Kubernetes is the leading container orchestration platform and has developed into the backbone technology for many organizations’ modern applications and infrastructure. As an open source project, “K8s” is also one of the largest success stories to ever emanate from the Cloud Native Computing Foundation (CNCF). In short, Kubernetes has revolutionized the way organizations deploy, manage, and scale applications.

How to Effortlessly Deploy Cribl Edge on Windows, Linux, and Kubernetes

Collecting and processing logs, metrics, and application data from endpoints have caused many ITOps and SecOps engineers to go gray sooner than they would have liked. Delivering observability data to its proper destination from Linux and Windows machines, apps, or microservices is way more difficult than it needs to be. We created Cribl Edge to save the rest of that beautiful head of hair of yours.

What Is AI Monitoring and Why Is It Important

Artificial intelligence (AI) has emerged as a transformative force, empowering businesses and software engineers to scale and push the boundaries of what was once thought impossible. However as AI is accepted in more professional spaces, the complexity of managing AI systems seems to grow. Monitoring AI usage has become a critical practice for organizations to ensure optimal performance, resource efficiency, and provide a seamless user experience.

Unveiling Splunk UBA 5.3: Power and Precision in One Package

In the face of an ever-evolving cybersecurity landscape, Splunk never rests. Today, we're ecstatic to share the release of Splunk User Behavior Analytics (UBA) 5.3, delivering power and precision in one package, and pushing the boundaries of what's possible in user and entity behavior analytics.

Getting _____________ for Less from Your Analytics Tools

Your analytics system of choice is probably pulling triple-duty for your enterprise–data collection, data storage, and its primary goal: analytics for monitoring, reporting and taking action. In this session we discuss considerations for various use cases, and why and how to use Cribl Stream to customize the processing and routing of various data sources to optimize, enrich, and route your data based on its content, value, and purpose.

Apica Acquires LOGIQ.AI to Revolutionize Observability

In the world of observability, having the right amount of data is key. For years Apica has led the way, utilizing synthetic monitoring to evaluate the performance of critical transactions and customer flows, ensuring businesses have important insight and lead time regarding potential issues.

Optimizing cloud resources and cost with APM metadata in Elastic Observability

Application performance monitoring (APM) is much more than capturing and tracking errors and stack traces. Today’s cloud-based businesses deploy applications across various regions and even cloud providers. So, harnessing the power of metadata provided by the Elastic APM agents becomes more critical. Leveraging the metadata, including crucial information like cloud region, provider, and machine type, allows us to track costs across the application stack.

Performance Testing: Types, Tools & Best Practices

To maximize the performance and value of your software apps, networks and systems, it’s critical to eliminate performance bottlenecks. Performance testing has become critical in every organization to reveal and fix performance bottlenecks, ensuring the best experience to end users. This article explains what performance testing is, its importance, and the various types of performance testing.

Cloud Analytics 101: Uses, Benefits and Platforms

Cloud analytics is the process of storing and analyzing data in the cloud and using it to extract actionable business insights. Simply one shade of data analytics, cloud analytics algorithms are applied to large data collections to identify patterns, predict future outcomes and produce other information useful to business decision-makers.

From Disruptions to Resilience: The Role of Splunk Observability in Business Continuity

In today's market, companies undergoing digital transformation require secure and reliable systems to meet customer demands, handle macroeconomic uncertainty and navigate new disruptions. Digital resilience is key to providing an uninterrupted customer experience and adapting to new operating models. Companies that prioritize digital resilience can proactively prevent major issues, absorb shocks to digital systems and accelerate transformations.

Managing your applications on Amazon ECS EC2-based clusters with Elastic Observability

In previous blogs, we explored how Elastic Observability can help you monitor various AWS services and analyze them effectively: One of the more heavily used AWS container services is Amazon ECS (Elastic Container Service). While there is a trend toward using Fargate to simplify the setup and management of ECS clusters, many users still prefer using Amazon ECS with EC2 instances.

What Does Real Time Mean?

Cindy works long hours managing a SecOps team at UltraCorp, Inc. Her team’s days are spent triaging alerts, managing incidents, and protecting the company from cyberattacks. The workload is immense, and her team relies on a popular SOAR platform to automate incident response including executing case management workflows that populate cases with relevant event data and enrichment with IOCs from their TIP, as well execute a playbook to block the source of the threat at the endpoint.

Configure Cribl Search to Explore and Catalog Petabytes of Data

If you’ve ever found yourself pondering the hidden treasures tucked away within thousands of files in Amazon S3, this is the perfect guide for you. In this blog post, we’re going to look at how you can use the Cribl Search fields feature to catalog and explore the fields in petabytes of data stored in Object Stores. In the Fields Tab within Cribl Search, all returned fields are categorized according to five different dimensions.

How to manually instrument iOS Applications with OpenTelemetry

In this tutorial, we dive into the practical application of OpenTelemetry in an iOS project built with SwiftUI. We demonstrate how to set up the OpenTelemetry SDK, generate spans, and send them to a configured OpenTelemetry collector. Our example application displays a random sentence every few seconds. Each sentence generation process is instrumented with OpenTelemetry spans, which include an attribute representing the word count of the sentence.

The Evolution of the Service Model In the Data Industry

Cribl’s Ed Bailey will lead a great discussion with nth degree’s Paul Stout and Scott Gray about the evolution of the service model from time and materials to outcome-based services. We will share our own stories about our experiences with services and how to make them better. Join the live stream for a fun discussion and come armed with suggestions for how to make your next services engagement better.

Logs Management & Correlating Logs with Traces in SigNoz

More about SigNoz: SigNoz - Monitor your applications and troubleshoot problems in your deployed applications, an open-source alternative to DataDog, New Relic, etc. Backed by Y Combinator. SigNoz helps developers monitor applications and troubleshoot problems in their deployed applications. SigNoz uses distributed tracing to gain visibility into your software stack. If you need any clarification or find something missing, feel free to raise a GitHub issue with the label documentation or reach out to us at the community slack channel.

The Top 15 Application Performance Metrics

Monitoring the key metrics of your application’s performance are essential to keep your software applications running smoothly as one of the key elements underpinning application performance monitoring. In this article, we will cover many of the key metrics that you should strongly consider monitoring to ensure that your next software engineering project remains fully performant.

Enhancements To Ingest Actions Improve Usability and Expand Searchability Wherever Your Data Lives

Splunk is happy to announce improvements to Ingest Actions in Splunk Enterprise 9.1 and the most recent Splunk Cloud Platform releases which enhance its performance and usability. We’ve seen amazing growth in the usage of Ingest Actions over the last 12 months and remain committed to prioritizing customer requests to better serve cost-saving, auditing, compliance, security and role-based access control (RBAC) use cases.

10 Best Dynatrace Alternatives [2023 Comparison]

Dynatrace has established itself as a prominent player in the field of application performance management, but given that Dynatrace is an expensive solution aimed at large enterprises, exploring your options is essential. This comprehensive article presents a handpicked selection of the top 10 Dynatrace alternatives, each offering distinct advantages and capabilities.

Follow Splunk Down a Guided Path to Resilience

The dynamic digital landscape brings risk and uncertainty for businesses in all industries. Cyber criminals use the advantages of time, money, and significant advances in technology to develop new tactics and techniques that help them evade overlooked vulnerabilities. Critical signals — like failures, errors, or outages — go unnoticed, leading to downtime and costing organizations hundreds of thousands of dollars.

How Qonto used Grafana Loki to build its network observability platform

Christophe is a self-taught engineer from France who specializes in site reliability engineering. He spends most of his time building systems with open-source technologies. In his free time, Christophe enjoys traveling and discovering new cultures, but he would also settle for a good book by the pool with a lemon sorbet.

Splashing into Data Lakes: The Reservoir of Observability

If you’re a systems engineer, SRE, or just someone with a love for tech buzzwords, you’ve likely heard about “data lakes”. Before we dive deep into this concept, let’s debunk the illusion: there aren’t any floaties or actual lakes involved! Instead, imagine a vast reservoir where you store loads and loads of raw data in its natural format. Now, pair this with the idea of observability and telemetry pipelines, and we have ourselves an engaging topic.

Managing Kubernetes Log Data at Scale | Civo Navigate NA 2023

In this talk by Matt Miller from Edge Deltas, we delve into the world of managing Kubernetes log data at a significant scale. Matt breaks down different strategies for handling log data, including Kubernetes native tools, persistent storage, open-source log collection, and using centralized log vendors. He explores the benefits and drawbacks of each approach, particularly focusing on the challenges of scalability and cost when dealing with large volumes of data. Matt also shares insights on how to optimize data management using an intelligent edge-first approach.

The Pleasure of Finding Things Out: Federated Search Across All Major Cloud Providers and Native Support for Amazon Security Lake

The newly released Cribl Search 4.2 brings enhancements that ease data management in today’s complex, cloud-centric environments. This update provides comprehensive compatibility with all major cloud providers – Amazon S3, Google Cloud Storage, and Azure Blob Storage. It also ushers in native support for Amazon Security Lake. In this blog post, we’ll examine how new dataset providers enhance the value that Cribl Search delivers, out of the box.

Understanding Linux Logs: Overview with Examples

Logging and program output are woven into the entire Linux system fabric. In Linux land, logs serve as a vital source of information about security incidents, application activities, and system events. Effectively managing your logs is crucial for troubleshooting and ensuring compliance. This article explores the importance of logging and the main types of logs before covering some helpful command line tools and other tips to help you manage Linux logs.

Data Retention Policy Guide

Data retention policy will become a major focus for CIOs in 2021. Here’s why: First, enterprise organizations are producing larger volumes of data than ever before and utilizing enterprise data across a wider range of business processes and applications. To maximize its value, this data must be managed effectively throughout its entire life cycle - from collection and storage, through to usage, archiving, and eventually deletion.

How to Monitor the Performance of Mobile-Friendly Websites

Mobile-friendly websites are a must. We are all using mobile devices more and more to access information and perform all kinds of work and tasks – shopping, banking, communication, dating, etc. Needless to say, if you operate a website, you more likely want to ensure people accessing it using mobile devices – tablets, smartphones, etc. – have a great experience.

Leveraging Cribl as an Integral Part of Your M&A Strategy

One of the most exciting things about bringing products to market at Cribl is seeing customers continually find new ways to leverage them to help solve their data challenges. I recently spoke to a customer who described Cribl as the foundation of their data management strategy and a key part of their post-acquisition data engineering process. Let’s take a deeper look into how Cribl can help.

Mainframe Observability with Elastic and Kyndryl

As we navigate our fast-paced digital era, organizations across various industries are in constant pursuit of strategies for efficient monitoring, performance tuning, and continuous improvement of their services. Elastic® and Kyndryl have come together to offer a solution for Mainframe Observability, engineered with an emphasis on organizations that are heavily reliant on mainframes, including the financial services industry (FSI), healthcare, retail, and manufacturing sectors.

Cribl Makes Waves at Black Hat USA 2023, Unveils Strategic Partnership with Exabeam to Accelerate Technology Adoption for Customers

One of our core values at Cribl is Customers First, Always. These aren’t just buzzwords we use to sound customer friendly; it’s ingrained in our daily communication and workload. Without our customers, we wouldn’t exist. One of the ways we’ve upheld this value is to seek out strategic partnerships with other companies aligned with our customers’ needs – both present and future.

DataDog Flex Logs vs Coralogix Remote Query

While Coralogix Remote Query is a solution to constant reingestion of logs, there are few other options today that also offer customers the ability to query unindexed log data. For instance, DataDog has recently introduced Flex Logs to enable their customers to store logs in a lower cost storage tier. Let’s go over the differences between Coralogix Remote Query vs Flex Logs and see how DataDog compares. Get a strong full-stack observability platform to scale your organization now.

ML-Powered Assistance for Adaptive Thresholding in ITSI

Adaptive thresholding in Splunk IT Service Intelligence (ITSI) is a useful capability for key performance indicator (KPI) monitoring. It allows thresholds to be updated at a regular interval depending on how the values of KPIs change over time. Adaptive thresholding has many parameters through which users can customize its behavior, including time policies, algorithms and thresholds.

IT Event Correlation: Software, Techniques and Benefits

IT event correlation is the process of analyzing IT infrastructure events and identifying relationships between them to detect problems and uncover their root cause. Using an event correlation tool can help organizations monitor their systems and applications more effectively while improving their uptime and performance.

Custom Java Instrumentation with OpenTelemetry

Discover how to gain unparalleled visibility and traceability into your applications with our latest video! As an expansion of our blog post, this tutorial dives into the world of Site Reliability Engineering and IT Operations, guiding you step-by-step through the process of implementing a plugin for the OpenTelemetry Java Agent using its Extensions framework.

Elasticsearch Vs OpenSearch | Comparing Elastic and AWS Search Engines

🔍 In this video, we will explore the differences between OpenSearch vs. Elasticsearch. We will look at the history of opensearch, compare performances, ecosystems, and much more. Brought to you by Sematext – your cloud monitoring and centralized logging solution.

Common API Vulnerabilities and How to Secure Them

Application programming interfaces (APIs) have become a critical part of almost every business. APIs are responsible for transferring information between systems within a company or to external companies. For example, when you log in to a website like Google or Facebook, an API processes your login credentials to verify they are correct.

Modeling and Unifying DevOps Data

“How can we turn our DevOps data into useful DevSecOps data? There is so much of it! It can come from anywhere! It’s in all sorts of different formats!” While these statements are all true, there are some similarities in different parts of the DevOps lifecycle that can be used to make sense of and unify all of that data. How can we bring order to this data chaos? The same way scientists study complex phenomena — by making a conceptual model of the data.

13 Best Cloud Cost Management Tools in 2023

Businesses are increasingly turning to cloud computing to drive innovation, scalability, and cost efficiencies. For many, managing cloud costs becomes a complex and daunting task, especially as organizations scale their cloud infrastructure and workloads. In turn, cloud cost management tools can help teams gain better visibility, control, and cost optimization of their cloud spending. These tools not only provide comprehensive solutions to track and analyze, they also optimize cloud expenses.

IDC Market Perspective published on the Elastic AI Assistant

IDC published a Market Perspective report discussing implementations to leverage Generative AI. The report calls out the Elastic AI Assistant, its value, and the functionality it provides. Of the various AI Assistants launched across the industry, many of them have not been made available to the broader practitioner ecosystem and therefore have not been tested. With Elastic AI Assistant, we’ve scaled out of that trend to provide working capabilities now.

Cloud Observability: Unlocking Performance, Cost, and Security in Your Environment

A robust observability strategy forms the backbone of a successful cloud environment. By understanding cloud observability and its benefits, businesses gain the ability to closely monitor and comprehend the health and performance of various systems, applications, and services in use. This becomes particularly critical in the context of cloud computing. The resources and services are hosted in the cloud and accessed through different tools and interfaces.

4 Node.js Logging libraries which make sophisticated logging simpler

Node.js logging, like any form of software instrumentation, isn’t an easy thing to get right. It takes time, effort, and a willingness to continue to iterate until a proper balance is struck. There are so many points to consider, including: Previously, here on the Loggly blog, I began exploring these questions in the context of three of the most popular web development languages: PHP, Python, and Ruby. But these aren’t the only popular languages in use today.

The Quixotic Expedition Into the Vastness of Edge Logs, Part 2: How to Use Cribl Search for Intrusion Detection

For today’s IT and security professionals, threats come in many forms – from external actors attempting to breach your network defenses, to internal threats like rogue employees or insecure configurations. These threats, if left undetected, can lead to serious consequences such as data loss, system downtime, and reputational damage. However, detecting these threats can be challenging, due to the sheer volume and complexity of data generated by today’s IT systems.

Integrating BindPlane Into Your Splunk Environment Part 2

Often it can be a challenge to collect data into a monitoring environment that does not natively support that data source. Bindplane can help solve this problem. As the Bindplane Agent is based on OpenTelemetry (and is also as freeform as possible), one can bring in data from disparate sources that are not easily supported by the Splunk Universal Forwarder.

Don't Drown in Your Data - Why you don't need a Data Lake

As a leader in Security Analytics, we at Elastic are often asked for our recommendations for architectures for long-term data analysis. And more often than not, the concept of Limitless Data is a novel idea. Other security analytics vendors, struggling to support long-term data retention and analysis, are perpetuating a myth that organizations have no option but to deploy a slow and unwieldy data lake (or swamp) to store data for long periods of time. Let’s bust this myth.

How to Manually Instrument .NET Applications with OpenTelemetry

Welcome to our deep-dive tutorial on manually instrumenting.NET applications with OpenTelemetry! In this comprehensive guide, we walk you through the process of adding OpenTelemetry to your.NET applications to help you better understand and optimize their performance. Whether you're an experienced.NET developer or just getting started, you'll find actionable insights and tips to improve your application monitoring and tracing capabilities.

How to Get Started with a Security Data Lake

Modern, data-driven enterprise SecOps teams use Security Information and Event Management (SIEM) software solutions to aggregate security logs, detect anomalies, hunt for security threats, and enable rapid response to security incidents. SIEMs enable accurate, near real-time detection of security threats, but today's SIEM solutions were never designed to handle the large amounts of security log data generated by modern organizations on a daily basis.

Dark Data: Discovery, Uses, and Benefits of Hidden Data

Dark data is all of the unused, unknown and untapped data across an organization. This data is generated as a result of users’ daily interactions online with countless devices and systems — everything from machine data to server log files to unstructured data derived from social media. Organizations may consider this data too old to provide value, incomplete or redundant, or limited by a format that can’t be accessed with available tools.

Data Lakes Explored: Benefits, Challenges, and Best Practices

A data lake is a data repository for terabytes or petabytes of raw data stored in its original format. The data can originate from a variety of data sources: IoT and sensor data, a simple file, or a binary large object (BLOB) such as a video, audio, image or multimedia file. Any manipulation of the data — to put it into a data pipeline and make it usable — is done when the data is extracted from the data lake.

Send your logs to multiple destinations with Datadog's managed Log Pipelines and Observability Pipelines

As your infrastructure and applications scale, so does the volume of your observability data. Managing a growing suite of tooling while balancing the need to mitigate costs, avoid vendor lock-in, and maintain data quality across an organization is becoming increasingly complex. With a variety of installed agents, log forwarders, and storage tools, the mechanisms you use to collect, transform, and route data should be able to evolve and adjust to your growth and meet the unique needs of your team.

Store and analyze high-volume logs efficiently with Flex Logs

The volume of logs that organizations collect from all over their systems is growing exponentially. Sources range from distributed infrastructure to data pipelines and APIs, and different types of logs demand different treatment. As a result, logs have become increasingly difficult to manage. Organizations must reconcile conflicting needs for long-term retention, rapid access, and cost-effective storage.

Leveraging Git for Cribl Stream Config: A Backup and Tracking Solution

Having your Cribl Stream instance connected to a remote git repo is a great way to have a backup of the cribl config. It also allows for easy tracking and viewing of all Cribl Stream config changes for improved accountability and auditing. Our Goal: Get Cribl configured with a remote Git repo and also configured with git signed commits. Git signed commits are a way of using cryptography to digitally add a signature to git commits.

BindPlane Agent Resiliency

A quick video about the resiliency of your BindPlane agent showing the parameters to tweak on destinations and best collector architecture to ensure you're not losing any data. About ObservIQ: observIQ brings clarity and control to our customer's existing observability chaos. How? Through an observability pipeline: a fast, powerful and intuitive orchestration engine built for the modern observability team. Our product is designed to help teams significantly reduce cost, simplify collection, and standardize their observability data.

Automatic log level detection reduces your cognitive load to identify anomalies at 3 am

Let’s face it, when that alert goes off at 2:58am, abruptly shaking you out of a deep slumber because of a high-priority issue hitting the application, you’re not 100% “on”. You need to shake the fog out of your head to focus on the urgent task of fixing the problem. This is where having the best log analytics tool can take on some of that cognitive load. Sumo Logic recently released new features specific to our Log Search queries that automatically detect log levels.

How to Tackle Spiraling Observability Costs

As today’s businesses increasingly rely on their digital services to drive revenue, the tolerance for software bugs, slow web experiences, crashed apps, and other digital service interruptions is next to zero. Developers and engineers bear the immense burden of quickly resolving production issues before they impact customer experience.

Crafting Prompt Sandwiches for Generative AI

Large Language Models (LLMs) can give notoriously inconsistent responses when asked the same question multiple times. For example, if you ask for help writing an Elasticsearch query, sometimes the generated query may be wrapped by an API call, even though we didn’t ask for it. This sometimes subtle, other times dramatic variability adds complexity when integrating generative AI into analyst workflows that expect specifically-formatted responses, like queries.

Dive Deeper into your Trace and Logs Data with Query Builder - Community Call Aug 1

This week for our community call we show our new Trace explorer with a GUI for creating queries, custom dashboards, and alert thresholds. Great participation from the community, thank you so much for participating. SigNoz - Monitor your applications and troubleshoot problems in your deployed applications, an open-source alternative to DataDog, New Relic, etc. Backed by Y Combinator. SigNoz helps developers monitor applications and troubleshoot problems in their deployed applications. SigNoz uses distributed tracing to gain visibility into your software stack.

Pipeline Efficiency: Best Practices for Optimizing your Data Pipeline

Data pipelines are the foundational support to any business analytics project. They are growing more critical than ever as companies are leaning on their insights to drive their business: 54% of enterprises said it was vital to their future business strategies. Data pipelines play a crucial role as they perform calculations and transformations used by analysts, data scientists, and business intelligence teams.

The Uphill Battle of Consolidating Security Platforms

A recently conducted survey of 51 CISOs and other security leaders a series of questions about the current demand for cybersecurity solutions, spending intentions, security posture strategies, tool preferences, and vendor consolidation expectations. While the report highlights the trends around platform consolidation over the short run, 82% of respondents stated they expect to increase the number of vendors in the next 2-3 years.

Announcing Easy Connect - The Fastest Path to Full Observability

Logz.io is excited to announce Easy Connect, which will enable our customers to go from zero to full observability in minutes. By automating service discovery and application instrumentation, Easy Connect provides nearly instant visibility into any component in your Kubernetes-based environment – from your infrastructure to your applications. Since applications have been monitored, collecting logs, metrics, and traces have often been siloed and complex.

Enable and use GKE Control plane logs

Are you having any issues with the control plane components in your GKE Cluster? Are you interested in gaining visibility into the control plane side of the cluster to troubleshoot the issues by yourself? Then GKE Control Plane Logs is a great way to gain insights on what's going on with your cluster. In this video, we provide a quick overview about Control Plane components and logs, and show how to enable control plane logs on the new and existing GKE clusters. Watch this video to learn how to use Control plane logs to troubleshoot webhook and control plane latency issues in GKE clusters.

Why we generate & collect logs: About the usability & cost of modern logging systems

Logs and log management have been around far longer than monitoring and it is easy to forget just how useful and essential they can be for modern observability. Most of you will know us for VictoriaMetrics, our open source time series database and monitoring solution. Metrics are our “thing”; but as engineers, we’ve had our fair share of frustrations in the past caused by modern logging systems that tend to create further complexity, rather than removing it.