Operations | Monitoring | ITSM | DevOps | Cloud

Analytics

Webinar Highlights: How Texas Instruments Uses InfluxDB

It’s back to school season, and oftentimes, that means people are purchasing TI-84 calculators for their kids. But did you know that Texas Instruments makes so much more than calculators? 😁 Michael Hinkle, a Probe Engineering and Manufacturing Supervisor at Texas Instruments, recently presented on “How Texas Instruments Uses InfluxDB to Upload Product Standards and to Improve Efficiencies”.

Elastic Workplace Search and Gmail: Unified search across all your content

As work from home has ballooned in 2020, virtual methods for communicating with colleagues have become more critical than ever. Same goes for all the useful productivity and collaboration tools at our disposal. The emerging downside is the difficulty of finding needed information among so many tools. Compounding the problem is the tendency for info to get siloed off by department.

Getting Started with Sending StatsD Metrics to Telegraf & InfluxDB

This tutorial will walk you through sending StatsD metrics to Telegraf. StatsD is a simple protocol for sending application metrics via UDP. These metrics can be sent to a Telegraf instance, where they are aggregated and periodically flushed to InfluxDB or other output sinks that you have configured. At the time of writing, we have 37 different output plugins supported.

Snapshot Lifecycle Management - Sept 9, 2020 Elastic Meetup

Snapshot Lifecycle Management (SLM) allows you to set up snapshot lifecycle policies to automate the timing, frequency, and retention of snapshots. In this talk we will cover the basics of registering a repository, setting up a snapshot policy and taking the test snapshots. If you are aware of the necessity of taking frequent snapshots for high availability/disaster recovery and you’ve been looking for a way to automate the whole process - then SLM is the right tool for you!

A Deep Dive into Machine Learning in Flux: Naive Bayes Classification

Machine learning — the practice of writing algorithms that improve automatically through experience — has become a buzzword nowadays that connotes to something otherworldly and on the bleeding edge of technology. I’m here to tell you while that may be true, getting started with machine learning doesn’t have to be hard!

Automate Elastic Cloud workflows using an SDK and Elasticsearch Service API

We recently announced the general availability of our Elasticsearch Service API. APIs help to automate tasks such as creating and scaling deployments, integrating with existing workflows, and testing. The Elasticsearch Service API supports the Open API Specification, which allows you to use tools like Swagger to generate software development kits (SDKs) in any programming language. You can import the API spec onto Postman and create a Postman Collection to create a test suite.

The Go client for Elasticsearch: Configuration and customization

In a previous blog, we saw that the seemingly simple job of an Elasticsearch client — moving data between the calling code and the cluster — is actually quite complicated under the hood. Naturally, as much as we try to make the default behaviour of the client optimal for the majority of scenarios, there are situations where you want to configure, customize, or enable/disable certain features.

Community Highlight: How InfluxDB Enables IoT Sensor Monitoring of Aquariums

I recently spoke with Jeremy White who is using InfluxDB to monitor his aquariums. By collecting IoT sensor data, he has been able to better understand his 200 gallon salt-water aquarium full of fish and coral. The entire project can be found on GitHub. Caitlin: Tell us about yourself and your career. Jeremy: I’m a Senior Network Automation Consultant at Network to Code, and my background is in networking engineering. Network to Code is an industry leader in network automation.

Bring new insights to your IP analytics with a global administrative layer in Elastic Maps

We love maps at Elastic. In the Elastic Stack, there is one core component of all data we visualize using maps: Location. Location can mean reporting real-time positions of fleet vehicles, using a geofence for limiting search results, gauging application performance metrics from a geographic area, or identifying security threats by attaching geographic coordinates to IP addresses.