I don't mean anything about trace/log. This is not about the size of the data. From the database perspective, IOPS is the key. You are asking for 4x IOPS for every metric index.
Sheng Wu 吴晟 Twitter, wusheng1108 dafang <13240156...@163.com> 于2022年11月28日周一 15:47写道: > > > > > I mean only metrics data,not trace data. Should metric data be much less than > trace? If yes, The data is only increase serval times.I have found our > metrics index(such as jvm_old_gc_count)is only 20G per day pre shard.It only > will be increase 4 times to 80G ,and trace index is more than 200G per shard > per day.So I thought it would be work properly. > > > > > 在 2022-11-28 15:40:42,"Sheng Wu" <wu.sheng.841...@gmail.com> 写道: > >AFAIK, there is no database that could live in 4 times large data I/O, > >even for ElasticSearch/OpenSearch. > > > >We are working on a new BanyanDB, which is much better in I/O, we hope > >one day we could provide this. But not for now. > > > > > >Sheng Wu 吴晟 > >Twitter, wusheng1108 > > > >dafang <13240156...@163.com> 于2022年11月28日周一 15:02写道: > >> > >> Hi god wu: > >> Recently, we have a seconds metrics demand.And we don't want to save these > >> metrics into our db.So we want to do an intercept when the metrics > >> aggregation in the oap(At this time, there will be two data streams: 1. > >> Data written to es in minutes; 2. Write the second level data of other > >> storage).But we found the metrics are setted the timebucket in minute when > >> they are created.So we have no idea to extend it, and we need your help~ > >> > >> > >> 为了怕表达不清楚,我再用中文表述一下: > >> 最近,我们有一个按秒级收集监控数据的需求。但是我们不想把这些秒级的数据存在es中。所以我们想在数据聚合的位置进行拦截(这时会同时存在2个数据流:1.写入es的以分钟为维度的数据;2.写入其他存储的秒级的数据)。但是我们阅读源码发现,metrics在生成的时候,就将timebucket阉割到分钟维度了,我们后面就没有办法扩展了。现在没有思路了,现在需要你的帮助 > >> > >> > >> Yours > >> 2022.11.28