cryptoe opened a new issue, #16183:
URL: https://github.com/apache/druid/issues/16183

   # Description
   
   Apache Druid 29.0.1 is a patch release that fixes some issues in the 29.0.0 
release. See the [complete set of 
changes](https://github.com/apache/druid/milestone/55?closed=1) for additional 
details.
   
   ### Bug fixes
   
   * Added type verification for INSERT and REPLACE to validate that strings 
and string arrays aren't mixed 
[#15920](https://github.com/apache/druid/pull/15920)
   * [Concurrent 
replace](https://druid.apache.org/docs/latest/ingestion/concurrent-append-replace)
 now allows pending Peon segments to be upgraded using the Supervisor 
[#15995](https://github.com/apache/druid/pull/15995)
   * Changed the `targetDataSource` attribute to return a string containing the 
name of the datasource. This reverts the breaking change introduced in Druid 
29.0.0 for INSERT and REPLACE MSQ queries 
[#16004](https://github.com/apache/druid/pull/16004) 
[#16031](https://github.com/apache/druid/pull/16031)
   * Decreased the size of the distribution Docker image 
[#15968](https://github.com/apache/druid/pull/15968)
   * Fixed an issue with SQL-based ingestion where string inputs, such as from 
CSV, TSV, or string-value fields in JSON, are ingested as null values when they 
are typed as LONG or BIGINT [#15999](https://github.com/apache/druid/pull/15999)
   * Fixed an issue where a web console-generated Kafka supervisor spec has 
`flattenSpec` in the wrong location 
[#15946](https://github.com/apache/druid/pull/15946)
   * Fixed an issue with filters on expression virtual column indexes 
incorrectly considering values null in some cases for expressions which 
translate null values into not null values 
[#15959](https://github.com/apache/druid/pull/15959)
   * Fixed an issue where the data loader crashes if the incoming data can't be 
parsed [#15983](https://github.com/apache/druid/pull/15983)
   * Improved DOUBLE type detection in the web console 
[#15998](https://github.com/apache/druid/pull/15998)
   * Web console-generated queries now only set the context parameter 
`arrayIngestMode` to `array` when you explicitly opt in to use arrays 
[#15927](https://github.com/apache/druid/pull/15927)
   * The web console now displays the results of an MSQ query that writes to an 
external destination through the 
[`EXTERN`](https://druid.apache.org/docs/latest/multi-stage-query/reference#extern-function)
 function [#15969](https://github.com/apache/druid/pull/15969)
   
   ### Incompatible changes
   
   #### Changes to `targetDataSource` in EXPLAIN queries
   
   Druid 29.0.1 includes a breaking change that restores the behavior for 
`targetDataSource` to its 28.0.0 and earlier state, different from Druid 29.0.0 
and only 29.0.0. In 29.0.0, `targetDataSource` returns a JSON object that 
includes the datasource name. In all other versions, `targetDataSource` returns 
a string containing the name of the datasource.
   
   If you're upgrading from any version other than 29.0.0, there is no change 
in behavior.
   
   If you are upgrading from 29.0.0, this is an incompatible change.
   
   [#16004](https://github.com/apache/druid/pull/16004)
   
   ### Dependency updates
   
   * Updated PostgreSQL JDBC Driver version to 42.7.2 
[#15931](https://github.com/apache/druid/pull/15931)
   
   
   
   ## Credits
   
   Thanks to everyone who contributed to this release!
   
   @abhishekagarwal87
   @adarshsanjeev
   @AmatyaAvadhanula
   @clintropolis
   @cryptoe
   @dependabot[bot]
   @ektravel
   @gargvishesh
   @gianm
   @kgyrtkirk
   @LakshSingla
   @somu-imply
   @techdocsmith
   @vogievetsky
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to