The short answer is “yes”. ClickHouse has multiple mechanisms that allow freeing up disk space by removing old data. Each mechanism is aimed for different scenarios.
ClickHouse allows to automatically drop values when some condition happens. This condition is configured as an expression based on any columns, usually just static offset for any timestamp column.
The key advantage of this approach is that it does not need any external system to trigger, once TTL is configured, data removal happens automatically in background.
TTL can also be used to move data not only to /dev/null, but also between different storage systems, like from SSD to HDD.
More details on configuring TTL.
ClickHouse does not have real-time point deletes like in OLTP databases. The closest thing to them are mutations. They are issued as
ALTER ... DELETE or
ALTER ... UPDATE queries to distinguish from normal
UPDATE as they are asynchronous batch operations, not immediate modifications. The rest of syntax after
ALTER TABLE prefix is similar.
ALTER DELETE can be issued to flexibly remove old data. If you need to do it regularly, the main downside will be the need to have an external system to submit the query. There are also some performance considerations since mutation rewrite complete parts even there’s only a single row to be deleted.
This is the most common approach to make your system based on ClickHouse GDPR-compliant.
More details on mutations.
ALTER TABLE ... DROP PARTITION provides a cost-efficient way to drop a whole partition. It’s not that flexible and needs proper partitioning scheme configured on table creation, but still covers most common cases. Like mutations need to be executed from an external system for regular use.
More details on manipulating partitions.
It’s rather radical to drop all data from a table, but in some cases it might be exactly what you need.
More details on table truncation.