Hi all,
I’m a beginner with QuestDB, and I found it very interesting!!
I would like to understand if can solve issues in huge archiving data and performance quering
I have data file CSV with 230 Million of rows daily and more over the Timestamp I have values and a Name, composed like a path
Name MainFolder.SubFolder.SubSubFolder.Name
I’m using the structure below, I’m asking to the Community if I should add something (indexing or other) to get more performance in queries?
If your paths have a limit to how nested they are, it might be a good idea to unroll them into separate columns, with the higher paths as SYMBOLs (if there aren’t too many distinct values). Filtering on SYMBOL columns is much faster than on strings.
You can also try searching using LIKE on the single path column, and see how fast it is.
It is better to use VARCHAR rather than STRING if you are not using SYMBOL. This will half the amount of storage you need if your text is mostly ASCII.
Hi, thanks’ for your quick answer!
The most common queries are time based (daily, hourly) looking for on the MainFolder and Name (the First and Last substrings) made in Grafana
Some time ago, I had to change a field from symbol to varchar, because with bulk inserts questdb stopped inserting data after a number of rows that varied randomly from 150000 to 250000. Have you found anything about this?
PS: the field had all distinct values