<h3><u>#general</u></h3><br><strong>@sjeetsingh2801: </strong>Hi 
guys!<br><strong>@adrian.f.cole: </strong>Hi, question about ServiceManager.. 
is it a feature or bug to have multiple instances of the same role in bootstrap 
services? eg. 2 minions<br><strong>@samarth: </strong>Is there a tool / utility 
to read segment data . modify it and upload . I was looking to generate bulk 
data for query performance testing pinot . Thanks ..<br><strong>@daniel.kocot: 
</strong>@daniel.kocot has joined the channel<br><strong>@wu.yj0616: 
</strong>@wu.yj0616 has joined the channel<br><strong>@vanshchaudhary: 
</strong>@vanshchaudhary has joined the channel<br><strong>@bmrja: 
</strong>@bmrja has joined the channel<br><strong>@daniel.kocot: </strong>Hi 
there :wave:<br><strong>@syed.zeeshan.ahmed: </strong>@syed.zeeshan.ahmed has 
joined the channel<br><h3><u>#random</u></h3><br><strong>@daniel.kocot: 
</strong>@daniel.kocot has joined the channel<br><strong>@wu.yj0616: 
</strong>@wu.yj0616 has joined the channel<br><strong>@vanshchaudhary: 
</strong>@vanshchaudhary has joined the channel<br><strong>@bmrja: 
</strong>@bmrja has joined the channel<br><strong>@syed.zeeshan.ahmed: 
</strong>@syed.zeeshan.ahmed has joined the 
channel<br><h3><u>#feat-presto-connector</u></h3><br><strong>@vkryuchkov: 
</strong>@vkryuchkov has joined the 
channel<br><h3><u>#troubleshooting</u></h3><br><strong>@npawar: </strong>it 
works for both<br><strong>@npawar: </strong>have you put epochMinutes in your 
schema?<br><strong>@pradeepgv42: </strong>yeah it’s part of the 
schema<br><strong>@pradeepgv42: </strong>let me try with the latest 
code<br><strong>@pradeepgv42: </strong>Actually with the latest code it worked, 
I was using slightly older version from master<br><strong>@arici: </strong>When 
I use Hour() function like below.
```SELECT "timestamp",variant_id,sum(amount) FROM Sales WHERE operator_id = 1 
AND campaign_id = 1 GROUP BY Hour("timestamp"), variant_id```
I get the following error. It says 'timestamp' should appear in GROUP BY 
clause. But it already does.
```ProcessingException(errorCode:150, message:PQLParsingError:
org.apache.pinot.sql.parsers.SqlCompilationException: 'timestamp' should appear 
in GROUP BY clause.
        at 
org.apache.pinot.sql.parsers.CalciteSqlParser.validateGroupByClause(CalciteSqlParser.java:177)
        at 
org.apache.pinot.sql.parsers.CalciteSqlParser.validate(CalciteSqlParser.java:114)
        at 
org.apache.pinot.sql.parsers.CalciteSqlParser.queryRewrite(CalciteSqlParser.java:364)
        at 
org.apache.pinot.sql.parsers.CalciteSqlParser.compileCalciteSqlToPinotQuery(CalciteSqlParser.java:338)
        at 
org.apache.pinot.sql.parsers.CalciteSqlParser.compileToPinotQuery(CalciteSqlParser.java:104)
        at 
org.apache.pinot.sql.parsers.CalciteSqlCompiler.compileToBrokerRequest(CalciteSqlCompiler.java:33)
        at 
org.apache.pinot.controller.api.resources.PinotQueryResource.getQueryResponse(PinotQueryResource.java:158)
        at 
org.apache.pinot.controller.api.resources.PinotQueryResource.handlePostSql(PinotQueryResource.java:131)
        at sun.reflect.GeneratedMethodAccessor32.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
org.glassfish.jersey.server.model.internal.ResourceMethodInvocationHandlerFactory.lambda$static$0(ResourceMethodInvocationHandlerFactory.java:52)
        at 
org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher$1.run(AbstractJavaResourceMethodDispatcher.java:124)
        at 
org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.invoke(AbstractJavaResourceMethodDispatcher.java:167))```<br><strong>@g.kishore:
 </strong>It’s not valid sql. You are selecting a column but grouping by a 
function on that column<br><strong>@arici: </strong>Also when I get rid of sum 
aggregation it works perfectly. (Is this a valid sql?)
```SELECT "timestamp",variant_id FROM Sales WHERE operator_id = 1 AND 
campaign_id = 1 GROUP BY Hour("timestamp"), variant_id```<br><strong>@arici: 
</strong>When I bring in the sum aggregation, I get the error 
above.<br><strong>@yash.agarwal: </strong>How to achieve something similar to 
`local.directory.sequence.id=true` in `SparkSegmentGenerationJobRunner` 
?<br><strong>@elon.azoulay: </strong>Hi, great talk yesterday! Have a question 
about the new date time field spec. I created a simple test table with the 
schema below but it does not seem to be bucketing into seconds (i.e. 
granularity). I inserted a time value every 500 ms to verify and did not see 
that the `ts` column is bucketed - is there another name for the "bucketed" 
column?

Here is the schema:
```{
  "schemaName" : "myTable2",
  "dimensionFieldSpecs" : [ {
    "name" : "col1",
    "dataType" : "LONG",
    "defaultNullValue" : 0
  } ],
  "metricFieldSpecs" : [ {
    "name" : "m1",
    "dataType" : "LONG"
  } ],
  "dateTimeFieldSpecs" : [ {
    "name" : "ts",
    "dataType" : "LONG",
    "format" : "1:MILLISECONDS:EPOCH",
    "granularity" : "1:SECONDS"
  } ]
}```
And here is the data:
```"resultTable": {
        "dataSchema": {
            "columnDataTypes": ["LONG", "LONG", "LONG"],
            "columnNames": ["col1", "m1", "ts"]
        },
        "rows": [
            [0, 0, 1598431406131],
            [1, 1000, 1598431406634],
            [2, 2000, 1598431407134],
            [3, 3000, 1598431407634],
            [4, 4000, 1598431408134],
            [5, 5000, 1598431408634],
            [6, 6000, 1598431409134],
            [7, 7000, 1598431409634],
            [8, 8000, 1598431410134],
            [9, 9000, 1598431410634]
        ]
    },```<br><strong>@g.kishore: </strong>you need to use 
transformFunction<br><strong>@elon.azoulay: </strong>Thanks! I saw examples in 
the unit tests, would I be using one of the functions with the `bucketed` 
suffix or just the regular `toEpochSeconds` transform?<br><strong>@g.kishore: 
</strong>@npawar ^^<br><strong>@npawar: </strong>if you want to round to the 
nearest seconds, but still have it in millis, then use `"tsSeconds" : 
"round(ts, 1000)"` in your transform function<br><strong>@npawar: </strong>if 
you want to convert it to secondsSinceEpoch, then you can use 
toEpochSeconds<br><strong>@npawar: </strong>and set that in your table config 
like this: 
<https://u17000708.ct.sendgrid.net/ls/click?upn=1BiFF0-2FtVRazUn1cLzaiMdTeAXadp8BL3QinSdRtJdpjVdLs6pcP-2BmJYu0RGOJgjW4yUWessZqxMl9cgpZEQvZ61b7e-2FULIdDgh3zQyBB6rFDS3rtq9FjWH9nlZ0hZ99VBNms8zq1x7TdCDGsaZg5i-2F97vZlpfEl6fip6rjl-2BUw-3DKTW8_vGLQYiKGfBLXsUt3KGBrxeq6BCTMpPOLROqAvDqBeTx1Zqjzg7h1GWrR9dxdr41nIlX5XgY4a7lmIKAYAmSJ8jxcdrosniuMabULDHSRR53N25lWt5Z31h4KoVp0i4rJrVHsZu6FU6E8TNzm0D2BD8msKdvc7zf0k0N5Bzp8r0ofAJyGHGjnAYV41H-2Fxd6BJablox7APRcEugKYHEM7-2Fvzn1ns-2FOgmeRyfH6r36NMAQ-3D><br><strong>@elon.azoulay:
 </strong>Thanks!<br><strong>@elon.azoulay: </strong>Did `streamConfigs` change 
to `ingestionConfig` in 0.5.0?<br><strong>@elon.azoulay: </strong>Or is it an 
additional config?<br><strong>@elon.azoulay: </strong>So the granularity by 
itself doesn't do anything by itself, a transform function is needed also, 
right?<br><strong>@npawar: </strong>it is additional config<br><strong>@npawar: 
</strong>correct about granularity<br><strong>@elon.azoulay: </strong>That 
really helps, thanks:) Also, great talk yesterday!<br><strong>@npawar: 
</strong>thank you! :slightly_smiling_face:<br><strong>@elon.azoulay: 
</strong>Sorry, another date time field spec question: can you have multiple 
date time columns bucketed on different granularities based on one source 
column?<br><h3><u>#lp-pinot-poc</u></h3><br><strong>@fx19880617: </strong>I got 
the docker build working, there is one missing thing about the istio 
context<br><strong>@fx19880617: 
</strong><https://u17000708.ct.sendgrid.net/ls/click?upn=1BiFF0-2FtVRazUn1cLzaiMSpQpf1vBNN7VUyMr5oI-2BtCAcH-2BA3iHH8tMnNcU0bwQJ7jx5-2B09UtwlYTMtvS0x-2BsWBHHckiDZf6ZXToOxKfdc-2Bq3yORTyU-2FdOxOeg-2FLAvytiO4o_vGLQYiKGfBLXsUt3KGBrxeq6BCTMpPOLROqAvDqBeTx1Zqjzg7h1GWrR9dxdr41nJBzgxN1syEs-2FjZcdBcX2k2p-2BoZjD3xoguKHwd-2FDD5tTFhIF7T50C7VXevEzr-2FjLkEVXjOnyxkQl8MNozVDRkHBubRLzMEAxJK-2FsaxliNIGLMiFIkEH6mo2UBvZCuAoFqdq7Zzz-2BpP-2Ff7l-2FOrxpCXDgU7a87kfNRRfwXoLetdPDE-3D><br>

Reply via email to