<h3><u>#general</u></h3><br><strong>@mailtobuchi: </strong>Quick question about 
Pinot query.
If this was the Pinot query result plan, does this mean `numSegmentsProcessed` 
segments were mem mapped?
```{
    "resultTable": {
        "dataSchema": {
            "columnDataTypes": ["BYTES"],
            "columnNames": ["id"]
        },
        "rows": [
            ["6a254bd3c853e950"]
        ]
    },
    "exceptions": [],
    "numServersQueried": 4,
    "numServersResponded": 4,
    "numSegmentsQueried": 1237,
    "numSegmentsProcessed": 1229,
    "numSegmentsMatched": 4,
    "numConsumingSegmentsQueried": 8,
    "numDocsScanned": 4,
    "numEntriesScannedInFilter": 4,
    "numEntriesScannedPostFilter": 4,
    "numGroupsLimitReached": false,
    "totalDocs": 265510367,
    "timeUsedMs": 32,
    "segmentStatistics": [],
    "traceInfo": {},
    "minConsumingFreshnessTimeMs": 1595297570989
}```
<br><strong>@mayanks: </strong>No, this is the number of segments the query had 
to process.<br><strong>@hiboss1: </strong>@hiboss1 has joined the 
channel<br><strong>@dlavoie: </strong>Wouldn't Pinot make an incredible 
datasource for Grafana?<br><strong>@pradeepgv42: </strong>@steotia Thanks a lot 
for enabling the TEXT_MATCH feature on dictionary encoded columns.
on a smaller table with ~25M rows, simple regexp_like query takes 178ms vs 
TEXT_MATCH takes ~30ms
This is pretty cool.<br><strong>@g.kishore: </strong>Amazing video by 
@kennybastani on Deploying Pinot on Kubernetes 
<https://u17000708.ct.sendgrid.net/ls/click?upn=1BiFF0-2FtVRazUn1cLzaiMc92bR9g-2BkGUUQX5IM7P9-2BAhHRVzOXTS92je0dEky-2B6erluPv4yRe5Qxf8-2BPzQrpHg-3D-3DJHPT_vGLQYiKGfBLXsUt3KGBrxeq6BCTMpPOLROqAvDqBeTybC1-2B-2FcUzX2RLE0WEiXpW8-2FoU2Y6JwxpYBGsUTMOfMOQsqm51Fyzpb3bLaYfh1TSyYryVJawHiPEsIw2FwM9lfYIO-2FfnyBg2faUY-2FGTgbUWpBy2hTlI03MzGgGB5mzTur8qzcpXO4hA0qczzwAyzOolfDvN60lMyZoH2Uda0Bgz2qIA-2BXBWWA94G1dOwatLvc-3D><br><strong>@rahulvinaykumar.chhap:
 </strong>Thanks for sharing :slightly_smiling_face:<br><strong>@sanjay: 
</strong>@sanjay has joined the 
channel<br><h3><u>#random</u></h3><br><strong>@hiboss1: </strong>@hiboss1 has 
joined the channel<br><strong>@sanjay: </strong>@sanjay has joined the 
channel<br><h3><u>#troubleshooting</u></h3><br><strong>@ankit.raj.singh: 
</strong>@ankit.raj.singh has joined the channel<br><strong>@elon.azoulay: 
</strong>We are about to upgrade to pinot-0.4.0 - do you recommend going to 
head or just cutting it at the 0.4.0 release commit?<br><strong>@elon.azoulay: 
</strong>Any notable config changes, or k8s changes we should be aware of? 
We're on pinot-0.3.0 now<br><strong>@damianoporta: </strong>Nooooo I have just 
upgraded my custom aggregation function :smile: did you change the 
API?<br><strong>@damianoporta: </strong>:joy:<br><strong>@g.kishore: 
</strong>@elon.azoulay I would go with 0.4.0 unless you need any feature in 
master<br><strong>@quietgolfer: </strong>Sorry, I think I've asked before (I 
lost my slack history).  Is there an easy way to have Pinot take the realtime 
inputs and automatically run data ingestion jobs to populate the offline 
tables?  Mostly checking to see if I can shortcut some work for a v1 
deliverable.  I assume there is probably a simple setup to output the kafka 
topic for 1 day, split the data and run batch ingestion 
jobs.<br><strong>@g.kishore: </strong>Yes, it’s doable but there is no such 
tool <br><strong>@g.kishore: </strong>You can download the real-time segments 
use Pinot segment reader to read multiple segments to generate a new offline 
segment and push it<br>

Reply via email to