3 minute read

Hi there!!! 👋

I have been working on some timeseries data for some time now and wanted to explore more on different timeseries databases out there.

So, there are a number of timeseries databases including influxdb, TimeScale, Prometheus, QuestDB, DataStax, Amazon timestream, dataPARC, etc.

A lot of other timeseries databases use and optimize postgreSQL in the background. Therefore, I decided to keep a simple template records of three most popular ones ones: influxDB, postgreSQL, and Prometheus.

So, what I will do is post templates for doing similar tasks: connection databases, and then storing a few fields including one column for the timestamp.


Here’s an example code for parsing 4/5 fields (including a timestamp) from JSON and storing the fields in InfluxDB in Python using the influxdb-client library:

from influxdb_client import InfluxDBClient
import json

# Load JSON data
data = '{"timestamp": "2022-01-01T00:00:00Z", "field1": 10, "field2": "value2", "field3": 3.14, "field4": true}'

# Parse JSON data
parsed_data = json.loads(data)

# Connect to InfluxDB
client = InfluxDBClient(url="http://localhost:8086", token="my-token", org="my-org")
write_api = client.write_api()

# Create InfluxDB point
point = {}
point["measurement"] = "my-measurement"
point["tags"] = {"tag1": "value1", "tag2": "value2"}
point["fields"] = {"field1": parsed_data["field1"], "field2": parsed_data["field2"], "field3": parsed_data["field3"], "field4": parsed_data["field4"]}
point["time"] = parsed_data["timestamp"]

# Write data to InfluxDB
write_api.write(bucket="my-bucket", record=point)

# Close connection to InfluxDB

This code assumes that you have already installed the influxdb-client library and have an InfluxDB instance running locally on port 8086 with a token and organization configured. It also assumes that the JSON data contains a timestamp field called “timestamp” and four other fields called “field1”, “field2”, “field3”, and “field4”.


Here’s an example code for parsing 4/5 fields (including a timestamp) from JSON and storing the fields in PostgreSQL using Python:

import psycopg2
import json
from datetime import datetime

# Create a connection to the PostgreSQL database
conn = psycopg2.connect(

# Create a cursor object to execute SQL commands
cur = conn.cursor()

# Load the JSON data
with open('data.json') as f:
    data = json.load(f)

# Parse the fields from the JSON data and store them in PostgreSQL
for item in data:
    timestamp = datetime.strptime(item['timestamp'], '%Y-%m-%d %H:%M:%S')
    field1 = item['field1']
    field2 = item['field2']
    field3 = item['field3']
    field4 = item['field4']

    # Execute an SQL INSERT statement to store the parsed fields in PostgreSQL
        "INSERT INTO your_table (timestamp, field1, field2, field3, field4) VALUES (%s, %s, %s, %s, %s)",
        (timestamp, field1, field2, field3, field4)

# Commit the changes and close the connection

Note that you’ll need to replace your_host, your_database, your_username, your_password, your_table, and data.json with the appropriate values for your setup.


Here’s an example code for parsing 4/5 fields (including a timestamp) from JSON and storing the fields in Prometheus using Python:

from prometheus_client import CollectorRegistry, Gauge, push_to_gateway
import json
import time

# Load JSON data from file
with open('data.json', 'r') as file:
    data = json.load(file)

# Create Prometheus registry and metric
registry = CollectorRegistry()
gauge = Gauge('metric_name', 'metric_help', ['field1', 'field2', 'field3', 'timestamp'], registry=registry)

# Parse data and push to Prometheus
for item in data:
    field1 = item['field1']
    field2 = item['field2']
    field3 = item['field3']
    timestamp = int(time.mktime(time.strptime(item['timestamp'], '%Y-%m-%d %H:%M:%S')))
    gauge.labels(field1=field1, field2=field2, field3=field3, timestamp=timestamp).set(item['value'])

# Push metrics to Prometheus server
push_to_gateway('localhost:9091', job='job_name', registry=registry)

Note that this code assumes that you have a Prometheus server running on localhost:9091 and that you have defined a job named job_name in your Prometheus configuration file.

So, that’s all for today! cheers!!!

Thank you for reading my blog post! 🙏

If you enjoyed it and would like to stay updated on my latest content and plans for next week, be sure to subscribe to my newsletter on Substack. 👇

Once a week, I’ll be sharing the latest weekly updates on my published articles, along with other news, content and resources. Enter your email below to subscribe and join the conversation for Free! ✍️

I am also writing on Medium. You can follow me here.

Leave a comment