|Aug 08, 2017||Capturing Perfmon Counters With Telegraf|
|Aug 07, 2017||InfluxDB and Annotations|
|May 24, 2017||Running InfluxDB as a service in Windows|
|May 17, 2017||Setting Up InfluxDb, Chronograf, and Grafana for the SqlServer Dev|
Updated: 2020-04-29broken image links removed
There are some beautiful ways to visualize time series data with the tools I’m going to go over. This post is purely focused on the initial setup and saving you some time there. In a future post, I’ll show how some of these tools can help you visualize your server performance in a powerful way, including taking metrics from multiple types of servers that be working with SQL Server, and combining the metrics when appropriate to give a full picture of performance.
A beautiful way to visualize performance across a variety of machines
It’s pretty epic to combine information across a variety of sources and be able to relate the metrics to the “big picture” that individual machine monitoring might fail to shed light on.
I started by running this quick powershell script to download the stable toolkit.
Once extracted, I moved the influx extracted subfolder into the InfluxDB folder to keep it clean. Now all the binaries rested in
C:\Influx\InfluxDB folder with no nesting folders.
I referenced the documentation for getting started with InfluxDB.
Setup Local InfluxDb
Started up the local influxdb binary.
Initializing the new database was simple as documented:
create database statty
Case SensitivityInfluxDB is case sensitive. Make sure to check your case if something isn’t working, such as use “DatabaseName” instead of use “databasename”
Also, if you get an error with access to the file, try running as admin.
More Enviromental Variable Fun
A simple fix to errors related to paths and the HOME variable these tools often need, per a Github issue, was to ensure the current path was available as a variable. I did this quickly with a simple batch file to launch the consoles as well as one option, as well as updated the Start-Process script to include a statement to set the env variable for the processes being started. This eliminated the issue. For more details see github issues
SET HOME=%~dp0 start influxd.exe start influx.exe
An additional snippet for launching the console version via a bat file:
set HOME=C:\influx cmd /k influx.exe -host "MyInfluxDbHost" -database "statty" -precision "s" -format column
Quick Start for Telegraf
Once you have this running you can take the telegraf binaries and run them on any other server to start capturing some default preset metrics. I launched with the following script and placed this in
C:\Influx directory to make it easy to access for future runs.
Edit the conf file to add some tags, change default sampling interval and more. I’ll post another article about setting up telegraf to run as a service in the future so search for more info
You can also apply the same bat file in the startup directory such as:
@REM alternative is using variable @REM set TELEGRAF_CONFIG_PATH=C:\telegraf\telegraf.conf start %~dp0telegraf.exe -config %~dp0telegraf.conf
One these metrics began to run, I ran Chronograf. This is Influx’s alternative to Grafana, another more mature product.
Upon loading and opening up the instance monitor, I found immediately that I was able to get some metrics from the defaults.
My preferred visualization tool, this was far more robust and well documented than Chronograf which has promise, but is a relatively new project.
When starting Grafana, you can run the following script. It creates a copy of the default ini to copy for the user to edit if not already there.
Once you open the localhost page, if you don’t see datasources in the left hand drop down, create an organization and ensure you are an admin, you’ll then see the option to add datasources. I simple pointed the page to InfluxDB console running on the server I had setup previously.
This is just a quick guide on getting started as I found a lot of little bumps in the road since the projects are written in
GO and not an easily run .NET project. Getting through this will hopefully give you a way to get started. I’ll blog a bit more soon on visualization of the metrics captured, some custom annotations to help make metrics come alive with real-time event notifications (like “load test started” and “build ended” etc). It’s a really promising solution for those who want some really nice flexibility in using perfmon and related metrics to visualize Windows and SQL Server performance.