https://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=24123
--- Comment #3 from Jonathan Druart <[email protected]> --- Created attachment 96160 --> https://bugs.koha-community.org/bugzilla3/attachment.cgi?id=96160&action=edit Bug 24123: Fix import of UTF-8 encoded MARC21 MARCXML using bulkmarcimport (elastic only) If elastic is used as search engine, the bulkmarcimport.pl will not handle correctly UTF-8 encoded MARCXML Koha::SearchEngine::Search->new uses a require statement to load the correct Search module. This is done l.257 of bulkmarcimport.pl: 257 my $searcher = Koha::SearchEngine::Search->new Koha::SearchEngine::Elasticsearch::Search will `use MARC::File::XML`, and so resets the arguments set before: 216 $MARC::File::XML::_load_args{BinaryEncoding} = 'utf-8'; 220 $MARC::File::XML::_load_args{RecordFormat} = $recordformat; An easy (but dirty) fix could be to move the declaration of my $searcher before in the script. The tricky (but correct) fix would be to remove the long standing "ugly hack follows" comment. This patch is the easy, and dirty, fix Test plan: Use the command line tool to import MARXCML records that contains unicode characters into Koha Something like `misc/migration_tools/bulkmarcimport.pl -biblios -file record.marcxml -m=MARCXML` Without this patch you will notice that unicode characters will not be displayed correctly -- You are receiving this mail because: You are watching all bug changes. You are the assignee for the bug. _______________________________________________ Koha-bugs mailing list [email protected] https://lists.koha-community.org/cgi-bin/mailman/listinfo/koha-bugs website : http://www.koha-community.org/ git : http://git.koha-community.org/ bugs : http://bugs.koha-community.org/
