X-Git-Url: http://koha-dev.rot13.org:8081/gitweb/?a=blobdiff_plain;f=misc%2Fmigration_tools%2Fbulkmarcimport.pl;h=4f738e896d709c22d00561886fd6ab01c0d14d5a;hb=c13d3f0043324ca9625128b25c9edba7b41ebd7f;hp=d6b8d401b936a26e1f2888471188edcee1904bd8;hpb=4e956892876626806437317e39451306372ba8c7;p=koha_fer diff --git a/misc/migration_tools/bulkmarcimport.pl b/misc/migration_tools/bulkmarcimport.pl index d6b8d401b9..4f738e896d 100755 --- a/misc/migration_tools/bulkmarcimport.pl +++ b/misc/migration_tools/bulkmarcimport.pl @@ -2,7 +2,8 @@ # Import an iso2709 file into Koha 3 use strict; -# use warnings; +use warnings; +#use diagnostics; BEGIN { # find Koha's Perl modules # test carefully before changing this @@ -19,17 +20,21 @@ use MARC::Charset; use C4::Context; use C4::Biblio; +use C4::Koha; +use C4::Debug; use C4::Charset; use C4::Items; use Unicode::Normalize; use Time::HiRes qw(gettimeofday); use Getopt::Long; -binmode(STDOUT, ":utf8"); - -use Getopt::Long; +use IO::File; +use Pod::Usage; -my ( $input_marc_file, $number) = ('',0); -my ($version, $delete, $test_parameter, $skip_marc8_conversion, $char_encoding, $verbose, $commit, $fk_off,$format); +binmode(STDOUT, ":utf8"); +my ( $input_marc_file, $number, $offset) = ('',0,0); +my ($version, $delete, $test_parameter, $skip_marc8_conversion, $char_encoding, $verbose, $commit, $fk_off,$format,$biblios,$authorities,$keepids,$match, $isbn_check, $logfile); +my ($sourcetag,$sourcesubfield,$idmapfl); +my $cleanisbn = 1; $|=1; @@ -37,6 +42,7 @@ GetOptions( 'commit:f' => \$commit, 'file:s' => \$input_marc_file, 'n:f' => \$number, + 'o|offset:f' => \$offset, 'h' => \$version, 'd' => \$delete, 't' => \$test_parameter, @@ -45,35 +51,31 @@ GetOptions( 'v:s' => \$verbose, 'fk' => \$fk_off, 'm:s' => \$format, + 'l:s' => \$logfile, + 'k|keepids:s' => \$keepids, + 'b|biblios' => \$biblios, + 'a|authorities' => \$authorities, + 'match=s@' => \$match, + 'i|isbn' => \$isbn_check, + 'x:s' => \$sourcetag, + 'y:s' => \$sourcesubfield, + 'idmap:s' => \$idmapfl, + 'cleanisbn!' => \$cleanisbn, ); +$biblios=!$authorities||$biblios; if ($version || ($input_marc_file eq '')) { - print < 2 ); + exit; +} + +if (defined $idmapfl) { + open(IDMAP,">$idmapfl") or die "cannot open $idmapfl \n"; +} + +if ((not defined $sourcesubfield) && (not defined $sourcetag)){ + $sourcetag="910"; + $sourcesubfield="a"; } my $dbh = C4::Context->dbh; @@ -89,10 +91,16 @@ if ($fk_off) { if ($delete) { - print "deleting biblios\n"; - $dbh->do("truncate biblio"); - $dbh->do("truncate biblioitems"); - $dbh->do("truncate items"); + if ($biblios){ + print "deleting biblios\n"; + $dbh->do("truncate biblio"); + $dbh->do("truncate biblioitems"); + $dbh->do("truncate items"); + } + else { + print "deleting authorities\n"; + $dbh->do("truncate auth_header"); + } $dbh->do("truncate zebraqueue"); } @@ -107,7 +115,8 @@ my $marcFlavour = C4::Context->preference('marcflavour') || 'MARC21'; print "Characteristic MARC flavour: $marcFlavour\n" if $verbose; my $starttime = gettimeofday; my $batch; -if ($format =~ /XML/i) { +my $fh = IO::File->new($input_marc_file); # don't let MARC::Batch open the file, as it applies the ':utf8' IO layer +if (defined $format && $format =~ /XML/i) { # ugly hack follows -- MARC::File::XML, when used by MARC::Batch, # appears to try to convert incoming XML records from MARC-8 # to UTF-8. Setting the BinaryEncoding key turns that off @@ -118,56 +127,215 @@ if ($format =~ /XML/i) { # extract the records, not using regexes to look # for .*. $MARC::File::XML::_load_args{BinaryEncoding} = 'utf-8'; - $batch = MARC::Batch->new( 'XML', $input_marc_file ); + my $recordformat= ($marcFlavour eq "MARC21"?"USMARC":uc($marcFlavour)); +#UNIMARC Authorities have a different way to manage encoding than UNIMARC biblios. + $recordformat=$recordformat."AUTH" if ($authorities and $marcFlavour ne "MARC21"); + $MARC::File::XML::_load_args{RecordFormat} = $recordformat; + $batch = MARC::Batch->new( 'XML', $fh ); } else { - $batch = MARC::Batch->new( 'USMARC', $input_marc_file ); + $batch = MARC::Batch->new( 'USMARC', $fh ); } $batch->warnings_off(); $batch->strict_off(); my $i=0; -my $commitnum = 50; +my $commitnum = $commit ? $commit : 50; -if ($commit) { -$commitnum = $commit; +# Skip file offset +if ( $offset ) { + print "Skipping file offset: $offset records\n"; + $batch->next() while ($offset--); +} +my ($tagid,$subfieldid); +if ($authorities){ + $tagid='001'; } +else { + ( $tagid, $subfieldid ) = + GetMarcFromKohaField( "biblio.biblionumber", '' ); + $tagid||="001"; +} + +# the SQL query to search on isbn +my $sth_isbn = $dbh->prepare("SELECT biblionumber,biblioitemnumber FROM biblioitems WHERE isbn=?"); $dbh->{AutoCommit} = 0; -RECORD: while ( my $record = $batch->next() ) { +my $loghandle; +if ($logfile){ + $loghandle= IO::File->new($logfile,"w") ; + print $loghandle "id;operation;status\n"; +} +RECORD: while ( ) { + my $record; + # get records + eval { $record = $batch->next() }; + if ( $@ ) { + print "Bad MARC record $i: $@ skipped\n"; + # FIXME - because MARC::Batch->next() combines grabbing the next + # blob and parsing it into one operation, a correctable condition + # such as a MARC-8 record claiming that it's UTF-8 can't be recovered + # from because we don't have access to the original blob. Note + # that the staging import can deal with this condition (via + # C4::Charset::MarcToUTF8Record) because it doesn't use MARC::Batch. + next; + } + # skip if we get an empty record (that is MARC valid, but will result in AddBiblio failure + last unless ( $record ); $i++; print "."; print "\r$i" unless $i % 100; - + + # transcode the record to UTF8 if needed & applicable. if ($record->encoding() eq 'MARC-8' and not $skip_marc8_conversion) { # FIXME update condition my ($guessed_charset, $charset_errors); - ($record, $guessed_charset, $charset_errors) = MarcToUTF8Record($record, $marcFlavour); + ($record, $guessed_charset, $charset_errors) = MarcToUTF8Record($record, $marcFlavour.(($authorities and $marcFlavour ne "MARC21")?'AUTH':'')); if ($guessed_charset eq 'failed') { warn "ERROR: failed to perform character conversion for record $i\n"; next RECORD; } } - + my $isbn; + # remove trailing - in isbn (only for biblios, of course) + if ($biblios && $cleanisbn) { + my $tag = $marcFlavour eq 'UNIMARC' ? '010' : '020'; + my $field = $record->field($tag); + my $isbn = $field && $field->subfield('a'); + if ( $isbn ) { + $isbn =~ s/-//g; + $field->update('a' => $isbn); + } + } + my $id; + # search for duplicates (based on Local-number) + if ($match){ + require C4::Search; + my $query=build_query($match,$record); + my $server=($authorities?'authorityserver':'biblioserver'); + my ($error, $results,$totalhits)=C4::Search::SimpleSearch( $query, 0, 3, [$server] ); + die "unable to search the database for duplicates : $error" if (defined $error); + #warn "$query $server : $totalhits"; + if ( @{$results} == 1 ){ + my $marcrecord = MARC::File::USMARC::decode($results->[0]); + $id=GetRecordId($marcrecord,$tagid,$subfieldid); + } + elsif ( @{$results} > 1){ + $debug && warn "more than one match for $query"; + } + else { + $debug && warn "nomatch for $query"; + } + } + my $originalid; + if ($keepids){ + $originalid=GetRecordId($record,$tagid,$subfieldid); + if ($originalid){ + my $storeidfield; + if (length($keepids)==3){ + $storeidfield=MARC::Field->new($keepids,$originalid); + } + else { + $storeidfield=MARC::Field->new(substr($keepids,0,3),"","",substr($keepids,3,1),$originalid); + } + $record->insert_fields_ordered($storeidfield); + $record->delete_field($record->field($tagid)); + } + } unless ($test_parameter) { - my ( $biblionumber, $biblioitemnumber, $itemnumbers_ref, $errors_ref ); - eval { ( $biblionumber, $biblioitemnumber ) = AddBiblio($record, '', { defer_marc_save => 1 }) }; - if ( $@ ) { - warn "ERROR: Adding biblio $biblionumber failed: $@\n"; - next RECORD; - } - eval { ( $itemnumbers_ref, $errors_ref ) = AddItemBatchFromMarc( $record, $biblionumber, $biblioitemnumber, '' ); }; - if ( $@ ) { - warn "ERROR: Adding items to bib $biblionumber failed: $@\n"; - # if we failed because of an exception, assume that - # the MARC columns in biblioitems were not set. - ModBiblioMarc( $record, $biblionumber, '' ); - next RECORD; - } - if ($#{ $errors_ref } > -1) { - report_item_errors($biblionumber, $errors_ref); + if ($authorities){ + use C4::AuthoritiesMarc; + my $authtypecode=GuessAuthTypeCode($record); + my $authid= ($id?$id:GuessAuthId($record)); + if ($authid && GetAuthority($authid)){ + ## Authority has an id and is in database : Replace + eval { ( $authid ) = ModAuthority($authid,$record, $authtypecode) }; + if ($@){ + warn "Problem with authority $authid Cannot Modify"; + printlog({id=>$originalid||$id||$authid, op=>"edit",status=>"ERROR"}) if ($logfile); + } + else{ + printlog({id=>$originalid||$id||$authid, op=>"edit",status=>"ok"}) if ($logfile); + } + } + elsif (defined $authid) { + ## An authid is defined but no authority in database : add + eval { ( $authid ) = AddAuthority($record,$authid, $authtypecode) }; + if ($@){ + warn "Problem with authority $authid Cannot Add ".$@; + printlog({id=>$originalid||$id||$authid, op=>"insert",status=>"ERROR"}) if ($logfile); + } + else{ + printlog({id=>$originalid||$id||$authid, op=>"insert",status=>"ok"}) if ($logfile); + } + } + else { + ## True insert in database + eval { ( $authid ) = AddAuthority($record,"", $authtypecode) }; + if ($@){ + warn "Problem with authority $authid Cannot Add".$@; + printlog({id=>$originalid||$id||$authid, op=>"insert",status=>"ERROR"}) if ($logfile); + } + else{ + printlog({id=>$originalid||$id||$authid, op=>"insert",status=>"ok"}) if ($logfile); + } + } + } + else { + my ( $biblionumber, $biblioitemnumber, $itemnumbers_ref, $errors_ref ); + $biblionumber = $id; + # check for duplicate, based on ISBN (skip it if we already have found a duplicate with match parameter + if (!$biblionumber && $isbn_check && $isbn) { + # warn "search ISBN : $isbn"; + $sth_isbn->execute($isbn); + ($biblionumber,$biblioitemnumber) = $sth_isbn->fetchrow; + } + if (defined $idmapfl) { + if ($sourcetag < "010"){ + if ($record->field($sourcetag)){ + my $source = $record->field($sourcetag)->data(); + printf(IDMAP "%s|%s\n",$source,$biblionumber); + } + } else { + my $source=$record->subfield($sourcetag,$sourcesubfield); + printf(IDMAP "%s|%s\n",$source,$biblionumber); + } + } + # create biblio, unless we already have it ( either match or isbn ) + if ($biblionumber) { + eval{$biblioitemnumber=GetBiblioData($biblionumber)->{biblioitemnumber};} + } + else + { + eval { ( $biblionumber, $biblioitemnumber ) = AddBiblio($record, '', { defer_marc_save => 1 }) }; + } + if ( $@ ) { + warn "ERROR: Adding biblio $biblionumber failed: $@\n"; + printlog({id=>$id||$originalid||$biblionumber, op=>"insert",status=>"ERROR"}) if ($logfile); + next RECORD; + } + else{ + printlog({id=>$id||$originalid||$biblionumber, op=>"insert",status=>"ok"}) if ($logfile); + } + eval { ( $itemnumbers_ref, $errors_ref ) = AddItemBatchFromMarc( $record, $biblionumber, $biblioitemnumber, '' ); }; + if ( $@ ) { + warn "ERROR: Adding items to bib $biblionumber failed: $@\n"; + printlog({id=>$id||$originalid||$biblionumber, op=>"insertitem",status=>"ERROR"}) if ($logfile); + # if we failed because of an exception, assume that + # the MARC columns in biblioitems were not set. + C4::Biblio::_strip_item_fields($record, ''); + ModBiblioMarc( $record, $biblionumber, '' ); + next RECORD; + } + else{ + C4::Biblio::_strip_item_fields($record, ''); + ModBiblioMarc( $record, $biblionumber, '' ); # need to call because of defer_marc_save + printlog({id=>$id||$originalid||$biblionumber, op=>"insert",status=>"ok"}) if ($logfile); + } + if ($#{ $errors_ref } > -1) { + report_item_errors($biblionumber, $errors_ref); + } } - $dbh->commit() if (0 == $i % $commitnum); } last if $i == $number; @@ -175,6 +343,7 @@ RECORD: while ( my $record = $batch->next() ) { $dbh->commit(); + if ($fk_off) { $dbh->do("SET FOREIGN_KEY_CHECKS = 1"); } @@ -184,9 +353,51 @@ $dbh->do("UPDATE systempreferences SET value=$CataloguingLog WHERE variable='Cat my $timeneeded = gettimeofday - $starttime; print "\n$i MARC records done in $timeneeded seconds\n"; - +if ($logfile){ + print $loghandle "file : $input_marc_file\n"; + print $loghandle "$i MARC records done in $timeneeded seconds\n"; + $loghandle->close; +} exit 0; +sub GetRecordId{ + my $marcrecord=shift; + my $tag=shift; + my $subfield=shift; + my $id; + if ($tag lt "010"){ + return $marcrecord->field($tag)->data() if $marcrecord->field($tag); + } + elsif ($subfield){ + if ($marcrecord->field($tag)){ + return $marcrecord->subfield($tag,$subfield); + } + } + return $id; +} +sub build_query { + my $match = shift; + my $record=shift; + my @searchstrings; + foreach my $matchingpoint (@$match){ + my $string = build_simplequery($matchingpoint,$record); + push @searchstrings,$string if (length($string)>0); + } + return join(" and ",@searchstrings); +} +sub build_simplequery { + my $element=shift; + my $record=shift; + my ($index,$recorddata)=split /,/,$element; + my ($tag,$subfields) =($1,$2) if ($recorddata=~/(\d{3})(.*)/); + my @searchstrings; + foreach my $field ($record->field($tag)){ + if (length($field->as_string("$subfields"))>0){ + push @searchstrings,"$index,wrdl=\"".$field->as_string("$subfields")."\""; + } + } + return join(" and ",@searchstrings); +} sub report_item_errors { my $biblionumber = shift; my $errors_ref = shift; @@ -199,3 +410,136 @@ sub report_item_errors { print $msg, "\n"; } } +sub printlog{ + my $logelements=shift; + print $loghandle join (";",@$logelements{qw}),"\n"; +} + + +=head1 NAME + +bulkmarcimport.pl - Import bibliographic/authority records into Koha + +=head1 USAGE + + $ export KOHA_CONF=/etc/koha.conf + $ perl misc/migration_tools/bulkmarcimport.pl -d -commit 1000 \\ + -file /home/jmf/koha.mrc -n 3000 + +=head1 WARNING + +Don't use this script before you've entered and checked your MARC parameters +tables twice (or more!). Otherwise, the import won't work correctly and you +will get invalid data. + +=head1 DESCRIPTION + +=over + +=item B<-h> + +This version/help screen + +=item B<-b, -biblios> + +Type of import: bibliographic records + +=item B<-a, -authorities> + +Type of import: authority records + +=item B<-file>=I + +The I to import + +=item B<-v> + +Verbose mode. 1 means "some infos", 2 means "MARC dumping" + +=item B<-fk> + +Turn off foreign key checks during import. + +=item B<-n>=I + +The I of records to import. If missing, all the file is imported + +=item B<-o, -offset>=I + +File offset before importing, ie I of records to skip. + +=item B<-commit>=I + +The I of records to wait before performing a 'commit' operation + +=item B<-l> + +File logs actions done for each record and their status into file + +=item B<-t> + +Test mode: parses the file, saying what he would do, but doing nothing. + +=item B<-s> + +Skip automatic conversion of MARC-8 to UTF-8. This option is provided for +debugging. + +=item B<-c>=I + +The I MARC flavour. At the moment, only I and +I are supported. MARC21 by default. + +=item B<-d> + +Delete EVERYTHING related to biblio in koha-DB before import. Tables: biblio, +biblioitems, items + +=item B<-m>=I + +Input file I: I or I (defaults to ISO2709) + +=item B<-k, -keepids>= + +Field store ids in I (usefull for authorities, where 001 contains the +authid for Koha, that can contain a very valuable info for authorities coming +from LOC or BNF. useless for biblios probably) + +=item B<-match>= + +I matchindex,fieldtomatch matchpoint to use to deduplicate fieldtomatch +can be either 001 to 999 or field and list of subfields as such 100abcde + +=item B<-i,-isbn> + +If set, a search will be done on isbn, and, if the same isbn is found, the +biblio is not added. It's another method to deduplicate. B<-match> & B<-isbn> +can be both set. + +=item B<-cleanisbn> + +Clean ISBN fields from entering biblio records, ie removes hyphens. By default, +ISBN are cleaned. --nocleanisbn will keep ISBN unchanged. + +=item B<-x>=I + +Source bib I for reporting the source bib number + +=item B<-y>=I + +Source I for reporting the source bib number + +=item B<-idmap>=I + +I for the koha bib and source id + +=item B<-keepids> + +Store ids in 009 (usefull for authorities, where 001 contains the authid for +Koha, that can contain a very valuable info for authorities coming from LOC or +BNF. useless for biblios probably) + +=back + +=cut +