summaryrefslogtreecommitdiffstats
path: root/backends
diff options
context:
space:
mode:
authorJoel Hans <joel@netdata.cloud>2019-11-13 09:00:00 -0700
committerGitHub <noreply@github.com>2019-11-13 09:00:00 -0700
commit4207f1797ebaf295aa954858d27bfbbeacaf5081 (patch)
treebd52b0ce39f6cf9547985349b1b15c2dea19268b /backends
parentb7d4ebabce4b116f0d494f9d2c6b049eaab2948a (diff)
TimescaleDB connection page (#7180)
* Added timescaledb page * Better title to fit tiny nav * GA link and some fixed grammar * Add reference in backends/README * Revert PostgreSQL change * Reverted on backends
Diffstat (limited to 'backends')
-rw-r--r--backends/README.md3
-rw-r--r--backends/TIMESCALE.md52
2 files changed, 55 insertions, 0 deletions
diff --git a/backends/README.md b/backends/README.md
index ac0847dcab..470544774b 100644
--- a/backends/README.md
+++ b/backends/README.md
@@ -39,6 +39,9 @@ So, although Netdata collects metrics every second, it can send to the backend s
metrics are labeled in the format, which is used by Netdata for the [plaintext prometheus
protocol](prometheus/). Notes on using the remote write backend are [here](prometheus/remote_write/).
+ - ****TimescaleDB** via [community-built connector](TIMESCALE.md) that takes JSON streams from a Netdata client
+ and writes them to a TimescaleDB table.
+
- **AWS Kinesis Data Streams**
metrics are sent to the service in `JSON` format.
diff --git a/backends/TIMESCALE.md b/backends/TIMESCALE.md
new file mode 100644
index 0000000000..6efb0729e9
--- /dev/null
+++ b/backends/TIMESCALE.md
@@ -0,0 +1,52 @@
+# Writing metrics to TimescaleDB
+
+Thanks to Netdata's community of developers and system administrators, and Mahlon Smith
+([GitHub](https://github.com/mahlonsmith)/[Website](http://www.martini.nu/)) in particular, Netdata now supports
+archiving metrics directly to TimescaleDB.
+
+What's TimescaleDB? Here's how their team defines the project on their [GitHub page](https://github.com/timescale/timescaledb):
+
+> TimescaleDB is an open-source database designed to make SQL scalable for time-series data. It is engineered up from
+> PostgreSQL, providing automatic partitioning across time and space (partitioning key), as well as full SQL support.
+
+## Quickstart
+
+To get started archiving metrics to TimescaleDB right away, check out Mahlon's [`netdata-timescale-relay`
+repository](https://github.com/mahlonsmith/netdata-timescale-relay) on GitHub.
+
+This small program takes JSON streams from a Netdata client and writes them to a PostgreSQL (aka TimescaleDB) table.
+You'll run this program in parallel with Netdata, and after a short [configuration
+process](https://github.com/mahlonsmith/netdata-timescale-relay#configuration), your metrics should start populating
+TimescaleDB.
+
+Finally, another member of Netdata's community has built a project that quickly launches Netdata, TimescaleDB, and
+Grafana in easy-to-manage Docker containers. Rune Juhl Jacobsen's
+[project](https://github.com/runejuhl/grafana-timescaledb) uses a `Makefile` to create everything, which makes it
+perferct for testing and experimentation.
+
+## Netdata&#8596;TimescaleDB in action
+
+Aside from creating incredible contributions to Netdata, Mahlon works at [LAIKA](https://www.laika.com/), an
+Oregon-based animation studio that's helped create acclaimed films like _Coraline_ and _Kubo and the Two Strings_.
+
+As part of his work to maintain the company's infrastructure of render farms, workstations, and virtual machines, he's
+using Netdata, `netdata-timescale-relay`, and TimescaleDB to store Netdata metrics alongside other data from other
+sources.
+
+> LAIKA is a long-time PostgreSQL user and added TimescaleDB to their infrastructure in 2018 to help manage and store
+> their IT metrics and time-series data. So far, the tool has been in production at LAIKA for over a year and helps them
+> with their use case of time-based logging, where they record over 8 million metrics an hour for netdata content alone.
+
+By archiving Netdata metrics to a backend like TimescaleDB, LAIKA can consolidate metrics data from distributed machines
+efficiently. Mahlon can then correlate Netdata metrics with other sources directly in TimescaleDB.
+
+And, because LAIKA will soon be storing years worth of Netdata metrics data in TimescaleDB, they can analyze long-term
+metrics as their films move from concept to final cut.
+
+Read the full blog post from LAIKA at the [TimescaleDB
+blog](https://blog.timescale.com/blog/writing-it-metrics-from-netdata-to-timescaledb/amp/).
+
+Thank you to Mahlon, Rune, TimescaleDB, and the members of the Netdata community that requested and then built this
+backend connection between Netdata and TimescaleDB!
+
+[![analytics](https://www.google-analytics.com/collect?v=1&aip=1&t=pageview&_s=1&ds=github&dr=https%3A%2F%2Fgithub.com%2Fnetdata%2Fnetdata&dl=https%3A%2F%2Fmy-netdata.io%2Fgithub%2Fbackends%2FTIMESCALE&_u=MAC~&cid=5792dfd7-8dc4-476b-af31-da2fdb9f93d2&tid=UA-64295674-3)](<>)