Time Nick Message 23:59 jcamins Run a detail record through Google's rich snippets tool. 23:58 dcook I don't recall seeing RDFa anywhere in Koha... 23:58 dcook Ummm, what? 23:58 dcook "The Evergreen and Koha integrated library systems now express their record details in the schema.org vocabulary out of the box using RDFa. " 23:48 dcook As with all things, I get the digests 23:47 dcook Ah, it probably just hasn't come in my mailbox yet 23:47 jcamins It's the last message in the HEADS UP thread. 23:46 dcook Well, except the system halt message 23:46 dcook I actually don't see anything in the mailing list. Hmm. *shrug* 23:46 dcook Thanks for the link 23:46 dcook Intriguing 23:45 jcamins http://stuff.coffeecode.net/www.loc.gov/marc/ 23:44 wahanui the mailing list is at http://koha-community.org/support/koha-mailing-lists/ 23:44 jcamins The mailing list. 23:44 dcook Where am I looking for code4lib? 23:43 jcamins Well, except for Z39.50. You're SOL there. 23:43 jcamins dcook: yeah, see code4lib. :) 23:42 dcook Hmm, anyone have any back-ups of the LoC website? :p 23:29 huginn 04Bug 10486: new feature, P5 - low, ---, jweaver, Needs Signoff , Allow external Z39.50 targets to be searched from the OPAC 23:29 huginn bag: The operation succeeded. 23:29 bag @later tell cait - http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=10486 updates added like requested :) 22:36 bag thanks wizzyrea 22:34 wizzyrea yep 22:33 rambuten koha-community you mean? 22:22 wizzyrea well it obviously *did* work as a denial of service heh. 22:19 wizzyrea Someone just slamming it with Joomla attacks. Silly people, that won't work it's wordpress. :P 22:02 rambuten http://www.koha-community.org/documentation <- works for me 22:01 wizzyrea oh look you're all checking to see if I'm telling the truth >.> 22:01 rambuten gmcharlt is in the weeds 21:59 wizzyrea that should be back 21:53 rangi gmcharlt: you around? 21:48 bag hey cait 21:45 wizzyrea yep 21:45 eythian wizzyrea: is that something you can poke at? 21:45 bag thanks for me too 21:45 cait1 hi bag 21:45 jcamins bag: Looks like it's down. 21:45 eythian bag: not working for me 21:44 huginn jcamins: I'll give you the answer just as soon as RDA is ready 21:44 jcamins @isitdown koha-community.org 21:43 bag anybody able to get that to load for them ^^^^ 21:43 wahanui manual is, like, at http://www.koha-community.org/documentation 21:43 bag manual? 21:31 cait1 :) 21:31 eythian cait1: guess I will :) 21:30 cait1 eythian: guess now you will have to come back here and buy him a beer? ;) 21:28 cait1 and order and receive an something on current master? 21:28 cait1 anyone feeling like they really want to do me a big favour? :) 21:24 cait1 restartedapache, now it works 21:24 cait1 rangi++ 21:21 wizzyrea bc if it's a gitified dev environment the logs will be where the packages keep them. 21:21 cait1 no, stlil haven't got around to try that 21:20 wizzyrea gitified? 21:20 cait1 so... does someone know why i have no log files appearing in my koha-dev/var/logs? :( 21:20 wizzyrea Just needed to know :) 21:20 wizzyrea yep that's fine 21:20 cait1 instead of adding another pref 21:20 cait1 for that purpose 21:20 cait1 hm i think the css here is intended 21:18 wizzyrea didn't know if that was a syspref or a hack your way around it thing./ 21:17 cait1 :) 21:17 wizzyrea ok cool ty 21:17 cait1 css 21:17 wizzyrea the authority magnifying glasses - can you turn those off? 21:16 rangi im going to do a dancer wrapper too 21:16 rangi NCIP needs to do both raw and http(s) to be a full implementation 21:14 rangi is for the mod_perl implementation 21:14 rangi then NCIPResponder.pm 21:14 rangi that runs raw 21:14 rangi which is called by bin/start_server.pl 21:14 rangi so, there is the NCIPServer.pm 21:13 rangi its another piece of the puzzle 21:11 wahanui que tal, eythian 21:11 eythian hi 20:58 rambuten http://www.wired.com/wiredenterprise/2013/09/gendarmerie_linux/ 20:58 magnuse rangi: did you ditch Dancer for NCIP, or is that another piece of the puzzle? 20:44 cait1 :) 20:44 cait1 glad you still got it 20:44 cait1 searches then... typos... :( 20:44 jcamins Yeah, that's Dialog. 20:44 wahanui searches are rendered client-side, particularly if you click the knowledge trail maps on the right. 20:44 cait1 searches 20:44 cait1 i remember carefully preparing searched than trying to be done really really fast with less possible searches from library school 20:43 cait1 possible other databases use their syntax or something? 20:43 cait1 it just sounded vaguely familiar 20:43 jcamins Weird, eh? 20:43 cait1 oh 20:43 jcamins Spending a small fortune per search somehow doesn't appeal to most people. 20:42 jcamins I only know of one thing that it is still used for regularly: citation searches for tenure applications. 20:42 jcamins Dialog _is_ the expensive database. 20:42 * cait1 should just look it up before asking stupid questions 20:41 rangi ? 20:41 cait1 was dialog that query langauge for expensive databases? 20:41 cait1 wow 20:41 jcamins Ooh, good news everyone! My library school has just added a course on Dialog to their schedule for the fall! 20:39 jeff well, i suppose "wasn't certain" isn't entirely accurate, but I'm not going to worry about being that precise. :P 20:38 jeff rangi: thanks for the clarification. that's what it looked like, but i wasn't certain. 20:38 rangi but its getting there 20:37 rangi my repo runs a little ahead, then gets merged 20:33 cait1 very cool :) 20:32 rangi its a collab between koha and evergreen people 20:32 rangi http://git.evergreen-ils.org/?p=working/NCIPServer.git 20:32 rangi jeff: thats just a mirror 19:56 jeff mtj: thanks. :-) 19:56 * jeff stars/watches 19:56 mtj ah, ok… i just spotted it on #koha yesterday :) 19:55 jeff mtj: I hadn't, thanks. It seems to be a currently outdated additional remote for the repo I linked earlier, http://git.evergreen-ils.org/?p=working/NCIPServer.git;a=summary 19:53 rambuten Does everybody have an NCIP dev/blog going on? 19:52 mtj jeff, have you spotted rangi's ncip repo on github? -> https://github.com/ranginui/ncip-server-chris 19:43 * jeff returns to making iNCIPit.cgi more useful 19:41 gaetan_B bye ! 19:40 * jeff sends sympathy gmcharlt's way -- good luck in the weeds 19:38 jeff ah, here: NCIP Responder Design: https://docs.google.com/document/d/1NuV47145SnEV1f-9BrY2d9jpX8Ij_VxK2a51lotgSsU/edit 19:37 jeff there's a developers-only list at the url gmcharlt posted (though others can supposedly lurk, as i understand it), and a document describing it at... 19:36 jeff There is an implementation of an NCIP 1.0 responder (intended for use with Evergreen and III's INN-REACH) here: https://github.com/jonscott/iNCIPit (with at least one more up to date fork available) and I believe Dyrcona and rangi and possibly others are collaborating on code here: http://git.evergreen-ils.org/?p=working/NCIPServer.git;a=summary 19:36 gmcharlt rambuten: briefly, because I'm in the weeds at the moment, you might want to consider joining http://lists.mvlc.org/listinfo/ncip-dev 19:36 cait1 there might be some changes in the program last minute, you neve rknow 19:36 rambuten hummm 19:36 cait1 rambuten: why not drop nancy a mail suggesting that? 19:35 rambuten gmcharlt: about? Any NCIP comments? 19:35 rambuten no, sadly 19:35 jcamins rambuten: are you staying for the hackfest? 19:35 mtompset rambuten: Sometimes less is more. :) 19:34 rambuten Wish KohaCon13 was having speed talks, I'd love to throw up some slides. 19:34 jeff Keep in mind that neither Evergreen nor Koha currently speak NCIP "out of the box". If you don't have a particular need for it, I'd say leave room/hooks/whatnot for future enhancement, but leave it at that. :-) 19:34 jcamins But I know dyrcona was working on NCIP a few weeks ago. 19:34 jcamins Not to say that jeff doesn't know about NCIP. He might. 19:34 rambuten OK, next time I see him online I'll ping him. 19:33 jcamins You should ask dyrcona about NCIP. 19:33 rambuten No, but "more is better", right? And there may be weird things out there that don't come to mind immediately where admins would prefer to use NCIP. 19:32 jeff Since both Evergreen and Koha do a good job of speaking SIP2, I'd almost suggest leaving it at that. Is there any particular reason you're thinking about supporting NCIP? Do you want to talk to other things that do talk NCIP and do not talk SIP2? 19:31 rambuten Public Access Computing - public Internet computers 19:30 jeff ah. just to be clear, can you define what you mean by "PAC" in this context? 19:30 rambuten SIP is a given, and probably NCIP will be included too. 19:30 rambuten Well, the PACs will authenticate to the Koha (and we hope to incorporate/test w/ Evergreen too) db server. 19:28 jeff What aspect of NCIP relates to your PAC development (or vice versa)? 19:27 oleonard Even in the open source world, people are often mum 19:24 jeff The other aspect to consider is that the NCIP standard defines a protocol of sorts, which is used to imeplement one or more Application Profiles. Some of those Application Profiles have various "flavors" in the wild, when it comes to interacting with specific software. 19:24 oleonard Can you say more? 19:24 rambuten We're 19:24 rambuten We funding a PAC development, and the dev had some questions. 19:24 oleonard Why do you ask rambuten? RDA is far from final but people are already "using" it 19:24 rambuten Yea, I was just looking at that site. 19:24 jeff rambuten: there are at least three currently published versions of the NCIP standard, and I suspect you could consider all of them finalized. see http://www.ncip.info/the-standard.html 19:24 jeff (sorry, snark is often not helpful) 19:24 jeff i'm not sure which of the last three words to mockingly place in quotation marks. 19:19 rambuten Is the NCIP protocol specification finalized? 18:53 cait1 permissions-- 16:41 oleonard *sigh* That was plenty of meeting. 16:18 kivilahtio__ jcamins: well I got your Koha INT_MAX :D 16:17 kivilahtio__ jcamins: ahhh :D 16:17 jcamins But Koha will. 16:17 jcamins No. 16:17 kivilahtio__ jcamins: do you really think MySQL goes crazy if you push it's integer boundaries? 16:17 kivilahtio__ jcamins: I am sure Mysql takes care of its boundary values. But I will write a warning about that to the help 16:16 jcamins You'll want to make *very* sure of that. 16:16 kivilahtio__ jcamins: we dont have that big id's 16:16 jcamins No, I mean in MySQL. 16:16 jcamins So if you import one record that has a 001 of 2147483646, you've just killed your database. 16:16 kivilahtio__ and force full reindex of Zebra 16:16 kivilahtio__ just need to realign all the foreign key references 16:15 kivilahtio__ jcamins: but the primary key can be compressed, it is a big script but doable when the time comes 16:15 jcamins 001 is often an arbitrarily large number. 16:15 kivilahtio__ jcamins: Agreed 16:15 jcamins The other danger I can see is primary key exhaustion. 16:13 kivilahtio__ jcamins: so others could benefit as well 16:13 kivilahtio__ jcamins: And I would rather have this work for us in the master Koha 16:13 kivilahtio__ jcamins: if the only problem is the purity of the 001, then it will work for us 16:13 kivilahtio__ jcamins: but it is good to know there are no Zebra related dangers 16:12 jcamins As I said, I have no objection to you adding this feature, I just think that it's likely to not work nearly as well as you imagine. 16:12 kivilahtio__ jcamins: Ill start working on that patch tomorrow 16:12 kivilahtio__ jcamins: naturally 16:12 kivilahtio__ jcamins: or in our case we collect the erroes in a huge log and manually clean up our databsae as we proceed with the migration 16:12 jcamins You'd have to make it optional, because data from Voyager, III, and most other ILSes that I have encountered does not consistently have 001 set correctly. 16:11 kivilahtio__ jcamins: this happens occasionally and I dont see an issue with that 16:11 kivilahtio__ jcamins: it can be manually migrated later 16:11 kivilahtio__ so the legacy record is not migrated 16:10 kivilahtio__ jcamins: well the biblionumber is primary key, so it cannot be dubplicated. Mysql deals with it 16:10 jcamins kivilahtio__: what happens when it turns out your previous system actually had a duplicate 001? 16:10 kivilahtio__ kf: Thakns for your help! 16:10 kf bye all 16:09 kf hm sorry have to leave 16:09 kivilahtio__ jcamins: why? we have staggered numbers, some records have been deleted on they fly, so our id sequence would be like 1001,1002,1004,1006,1007,1008,1009,1011 16:09 jcamins Itemnumber is of course unrelated. 16:09 jcamins If you have a single bad record, a single inconsistency, you are likely to trash your database. 16:08 kivilahtio__ jcamins: we use a strict integer, 0< 16:08 jcamins kivilahtio__: nothing, the problem is making the biblionumber match the 001. 16:08 kivilahtio__ jcamins: so what risks do you see in keeping the 001 matching the biblionumber and itemnumber? 16:08 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=9921 enhancement, P5 - low, ---, nunyo, Patch doesn't apply , Make it possible to force 001 = biblionumber 16:08 kf bug 9921 16:07 jcamins kivilahtio__: it is almost surely not easier. 16:07 kivilahtio__ and make this odd bugfix while at it 16:07 kivilahtio__ jcamins: I understand. But for me it is easier to just keep the 001 as a biblionumber :D 16:07 kf hm there is also a bug for that 16:07 jcamins I think making it possible to keep 001 as biblionumber would be nice, but not really necessary. 16:07 kivilahtio__ kf: I am only afraid of zebra 16:06 jeff j_wright++ -- thanks! (re: Plack pointers) 16:06 kf some people have voted for that bug - i am not a fan, but i just think meddling with the numbes is prone to create headache later 16:06 kivilahtio__ well within the databse id range 16:06 kivilahtio__ starting from 1001 -> 16:06 kivilahtio__ kf: yes 16:06 kivilahtio__ kf: I cant link holdins records to the bibliographic record 16:06 kf are your legacy ids strictly numeric? 16:05 kf what happens if koha assigns the next number? 16:05 kivilahtio__ programmatically it is not a problem, but I feel there are some sinister drawbacks to not using the in-DB primary key sequence? 16:05 kivilahtio__ kf: but what is the handicap for using our legacy id's? I have a solution in mind for that already and I think I could du that bug which was posted 16:04 kf as an example 16:04 kf kivilahtio__: the link above is a result list of articles 16:04 kivilahtio__ kf: ok 16:04 kf and they could be the same thing, but it doesn't matter for yourl inking, for your linking it's only important that you keep your 001 and $w subfields intact 16:03 kf kivilahtio__: I don't think there is a definitoin that your internal record number has to match the controlnumber 16:03 jcamins Whereas EG does a lot at the database level, Koha's bibliographic relationships are controlled only at the MARC level, in Zebra. 16:03 kf [off] https://hfjs.bsz-bw.de/cgi-bin/koha/opac-search.pl?idx=kw&op=and&idx=kw&op=and&idx=kw&limit=bib-level%3Aa&sort_by=relevance&do=Suche 16:02 kivilahtio__ jcamins: I hope I understood that better 16:02 kivilahtio__ kf: but why cant they be the same? they are bu definition the same thing? 16:02 jcamins Remember, Koha and EG treat records rather differently. 16:02 kf which is not the bilbionumber 16:02 kf the 001 of the host record 16:02 jcamins The 001 field. 16:02 kf linkings from 001 to $w subfields 16:02 kivilahtio__ jcamins: so where does 773w point in koha? 16:02 kf we do exactly that in koha 16:02 jcamins Right, and that's not biblionumber. 16:01 kivilahtio__ but our legacy data has that linkage 16:01 jcamins 773$w->001 doesn't use biblionumber. 16:01 kf kivilahtio__: 001 has nothing to do with biblionumber 16:01 kf and if you use an sql script... i would carry the number you need for linking in another custom field and adapt the sql accordingly 16:01 kivilahtio__ we cant change the database id 16:01 kivilahtio__ jcamins: well we have component parts with links from 773w -> 001 of the parent 16:00 jcamins I'd prepare MARC records that included the items prior to the import. 16:00 kf and import them with the marc records 16:00 kf well for other migrations, for items + biblio we attach the items as 952 15:59 kf we store the former id in a custom field and then look it up 15:59 kf that#s what we do 15:59 kf you could write a script that looked up the number 15:59 kivilahtio__ kf: just too much reworking our migration tool 15:59 kivilahtio__ kf: I cant migrate holdings with the marc records 15:59 kf you can have a biblionumber, of course you need one, but we let koha set them. 15:59 jcamins That sounds like a one-time task. 15:58 jcamins kivilahtio__: what do you mean by "link your items"? 15:58 j_wright jeff: XYZ is, rename app.psgi to app.cgi, change the shebang line to #!/usr/bin/env plackup or the three lines from the POD 15:58 kivilahtio__ kf: if I can't have a biblionumber set in biblio-table, how can I link my items, when they have the biblionumber in them pointing to the parent record? 15:57 kf i am still not sure it's a good idea 15:57 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=6113 enhancement, P5 - low, ---, gmcharlt, ASSIGNED , enhancement to keep previous ids 15:57 kf see bug 6113 15:56 kf the 999 are special 15:56 kivilahtio__ kf: But that is what the bulkmarcimport uses from "Koha to MARC mapping" 15:56 kf kivilahtio__: it's not made to do what you want 15:56 kf kivilahtio__: koha gives the bilbionumber, it's not supposed that you bring your own 15:55 j_wright jeff: Plack::Handler::CGI 15:55 kf kivilahtio__: there is probably not 15:54 kivilahtio__ Issues is that the biblionumber is correctly brought up to the database accessor function but it is not stored in the DB, but a new one is generated from the database primary key sequence 15:54 kivilahtio__ anyway, I'm going home 15:53 kivilahtio__ biblionumber is picked fine up until that function, where nothing is done upon the biblionumber? I mean there should be a way to map the biblionumber? 15:52 kivilahtio__ hey there is something wrong with the bulkmarcimport.pl script. There seems to be no way of pushing the old biblio id to the biblionumber! The marc mappings don't do anything since the database accessor functions don't even touch the biblionumber column! in C4::Biblio::_koha_add_biblio() 15:51 jcamins I know very little about native-PSGI applications. 15:51 jcamins There may be something. 15:50 jeff Thanks. It was worth a shot. I had this dream of writing a script that talks PSGI and then in the README says "if you must run this under a CGI environment, do XYZ" 15:50 kf bye all 15:49 jcamins jeff: I am not. 15:46 jeff jcamins: That last one isn't Koha-specific at all, sorry. :-) 15:46 jeff jcamins: I understand that Plack has support for running (most, with some assumptions/fixes) CGI-style scripts under PSGI using Plack::App::CGIBin. Are you aware of anything that does the reverse -- takes a script that speaks PSGI and allows it to run under a CGI environment? 15:41 jcamins However, Koha under Plack is still a bit brittle. 15:39 jcamins The only things that absolutely don't work under Plack are database export, and the progress bar for importing records. 15:37 jcamins At this point I think we've addressed all the variable scoping issues, mostly with terrible bandaids. 15:36 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=7172 major, P1 - high, ---, paul.poulain, NEW , Omnibus for Plack variable scoping problems 15:36 jeff jcamins: http://wiki.koha-community.org/wiki/Plack and the referenced bug 7172 both seem somewhat dormant. Is that because the work's done, or because there hasn't been much progress to note lately? 15:36 jcamins jeff: most things work under Plack. 15:35 rambuten no speed talks @kohacon13 ? 15:35 jcamins jeff: probably one of the best you're going to get, anyway. 15:32 jeff kivilahtio__: greetings! 15:31 jeff jcamins: for starters, the "coding for plack" idea on the hackfest wiki page has me wondering what the current state of plack is in koha? 15:31 kivilahtio__ jeff :) 15:31 jeff jcamins: Would you consider yourself a good person whose brain I could pick about Koha and Plack? 15:21 rambuten will/well 15:21 rambuten I'm gratified to see that coffee break sessions are will represented. I wonder if I could offer to bring/have purchased some better-than-your-average-bear coffee, like Eight O' Clock 15:21 kf or... 15:20 kf you could talk about the awesome query parser 15:20 kf :) 15:20 jcamins I am, yes. 15:20 kf jcamins: i think oyu are around the first day of the hackfest? 15:19 wahanui hackfest13 is http://wiki.koha-community.org/wiki/Kohacon13/Hackfest 15:19 kf hackfest13? 15:19 kf hackfest13 is http://wiki.koha-community.org/wiki/Kohacon13/Hackfest 15:19 kf I will try to send out an email about the hackfest planning tomorrow or so I think 15:19 wahanui hackfest is awesome! Group motivated koha hacking is the best koha hacking :) 15:19 kf hackfest? 15:18 jcamins That's the one. 15:18 rambuten http://wiki.koha-community.org/wiki/KohaCon13_Program 15:17 jcamins Yeah, the KohaCon13 pages are kind of unusable. There's one somewhere, though. 15:17 rambuten it's not exactly jumping out at me on any of the pages I see 15:16 jcamins There is. 15:16 jcamins For the conference? 15:15 rambuten apparently no presentation schedule yet? 15:14 jcamins We could probably add it back below "Community Resources." 15:14 rambuten kinda annoying to have to do a google search to find something on a given site. 15:12 jcamins I think so, yeah. 15:12 rambuten ? 15:12 rambuten didn't the old koha-community have a search field on the main page 14:47 kivilahtio__ bah ui misunderstanding 14:45 kivilahtio__ jcamins: ok I just realized my logical mistake 14:45 kivilahtio__ jcamins: well it takes time to rewite the scripts. I was hoping to reuse some Koha parts which use the biblionumber. 14:44 jcamins kivilahtio__: how often do you merge bibs and items? 14:44 kivilahtio__ I just save my new mappings but nothing happens... 14:43 kivilahtio__ but this might be really hard when merging biblios and items 14:43 kivilahtio__ kf: lol I cant set my mappings back to the defaults! Ok I'll try the $999 route 14:41 kf kivilahtio__: we never preserve old ids in migrations, but uor 001 are unique from a union catalog. We only store the ids from former systems in some fields so looking them up would be possible 14:40 jcamins rambuten: nope. 14:40 kf kivilahtio__: messing with the ids will bring you all kinds of headache later 14:40 rambuten jcamins: you didn't kick ban, did you? 14:38 rambuten I never have problems getting the nick, but it's no big deal 14:37 jcamins Oh. 14:37 rambuten somebody has copyrighted it, but they never seem to be online 14:37 jcamins rambuten: if it's registered, you can use /ghost rambutan, I think. 14:37 jcamins rambuten: rambutan may not have left the network yet. 14:37 kivilahtio__ rambuten: maybe wait a bit? 14:37 kivilahtio__ rambuten: maybe it is still reserved? 14:37 rambuten strangely it won't let me assume that nick now 14:36 jcamins Done. 14:36 jcamins You can try that. I'm actually not sure how well it will work. 14:36 rambuten jcamins: can you kick rambutan for me? 14:36 kivilahtio__ so if I want to keep our existing dabase id's I just need to sucker it up and migrate them in the $999? 14:36 jcamins That's what I thought. That's why I leave biblionumber in 999$c and biblioitemnumber in 999$d. 14:35 kivilahtio__ jcamins: wow I don't want to do that :D 14:35 kivilahtio__ default id is in $999, but we have it in $001 14:35 jcamins You're also going to have to use a custom Zebra configuration. 14:35 kivilahtio__ jcamins: hmm, I thought I understood that much from reading the bulmarcimport.pl code 14:35 jcamins Personally, I wouldn't do that. 14:34 kivilahtio__ jcamins: ERROR: Adding biblio failed: No biblionumber tag for framework "" 14:33 jcamins kivilahtio__: you should not be mapping biblioitem.biblionumber, and biblioitemnumber != biblioitem.biblionumber 14:31 kivilahtio__ I would like to preserve our old database id sitting in $001 14:31 kivilahtio__ hey I cannot duplicate the field 001 in "Administration" -> "Koha to Marc mapping". I need to explicitly define the field 001 to biblio.biblionumber, biblioitem.biblionumber and biblioitem.biblioitemnumber but I can apparently only assign it once 14:26 banana wizzyrea++ -> "we would only include other Open Source projects" 13:53 Barrc Of course it does........half asleep here, thanks! 13:53 jcamins Barrc: the second number is the one that marks stable/development. 13:51 Barrc Hey all - are odd Koha version numbers still considered unstable/dev releases? I ask cause I just ran a package install and then koha-install-log say "KOHA_INSTALLED_VERSION=3.12.05.000" 13:34 smeagol Hi all, does anyone know if localuse syspref counts items that are checked in to koha to trigger a HOLD to waiting status? thanks, as always 13:30 mtompset 12 hours reindexing... and still not done. Oh the pain! 13:21 marcelr cya 13:16 marcelr fine finer finest 13:15 magnuse in other news, i would hesitate to call addbiblio.pl "fine"... 13:15 magnuse yay! 13:14 kf yep 13:11 marcelr so that is fine 13:11 marcelr code from Installer.pm 13:11 marcelr kf: i did now 13:10 marcelr my $request = $self->{'dbh'}->prepare( "INSERT IGNORE INTO `systempreferences` (variable,value,explanation,options,type) VALUES('marcflavour','$marc_cleaned','Define global MARC flavor (MARC21, UNIMARC or NORMARC) used for character encoding','MARC21|UNIMARC|NORMARC','Choice');" ); 13:09 kf and hi marcelr 13:09 kf sorry was afk until now 13:09 kf guess you checked :) 13:09 marcelr $installer->set_marcflavour_syspref($marcflavour); ? 13:08 marcelr hi kf 13:08 kf not sure if the webinstaller creates the syspref or only updates it 13:07 kf hm 13:00 magnuse marcelr: me too :-/ 13:00 marcelr magnuse: after you please.. i have so much to fix already :) 12:59 kivilahtio__ magnuse: but that is a good idea 12:59 magnuse marcelr: it's all yours if you want to fix it ;-) 12:59 kivilahtio__ magnuse: our barcodes are not 100% reliable 12:59 magnuse kivilahtio__: use barcodes for the matching instead? 12:58 kivilahtio__ we migrate it all 12:58 kivilahtio__ and fines 12:57 kivilahtio__ magnuse: so we can target our circulations and holds to our items 12:57 magnuse kivilahtio__: why would you need or want to preserve the itemnumbers? 12:56 marcelr so in a specific case it should not appear 12:56 kivilahtio__ or maybe there is a more stylish alternative for migrating items using the Koha API? 12:56 marcelr magnuse: no sorry there is an exit in between 12:55 kivilahtio__ just installing my git environment and figuring stuff out 12:55 kivilahtio__ Hi I was migrating items to Koha and realized that the itemnumber was not preserved. I used C4::Items::_koha_new_item() to migrate items. Would it be acceptable if I extended that function to push the itemnumber to DB as well if it is present? 12:55 marcelr without the usmarc check 12:54 marcelr in the code more below the first line 12:54 marcelr magnuse: the situation seems to corrected by a later line: $template->param( title => $record->title() ) if ( $record ne "-1" ); 12:51 magnuse been like that since 2007, i think :-) 12:50 marcelr apparently, it does not hurt much 12:50 marcelr good catch 12:50 marcelr somewhere a title is not filled in the template 12:47 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=10975 normal, P5 - low, ---, gmcharlt, NEW , addbiblio.pl checks if marcflavour equals "usmarc" 12:47 magnuse Bug 10975 12:47 magnuse aww... 12:46 marcelr holmarc never made it ;) 12:46 magnuse w00t! 12:44 marcelr normarc is one of the big three :) 12:44 marcelr uppercase 12:42 marcelr marc21 12:42 magnuse but at least marcflavour will never be "usmarc", right? 12:40 magnuse thanks marcelr 12:39 magnuse ah yes, that could be it 12:39 marcelr it should not be in that file 12:38 marcelr magnuse: isn't that a specific choice in webinstaller? 12:37 magnuse is it just me, or is marcflavour missing from the new, alphabetically sorted sysprefs.sql? 12:25 magnuse ooh, shiny new buttons on the cataloguing page! 12:20 magnuse kia ora oleonard 12:20 oleonard Hi #koha 12:17 * magnuse waves again 11:20 paxed that should really complain something else if the db is "missing" ...) 11:19 paxed ergh. right, forgot port 8080, derp. 11:07 paxed any dev got some time to help me resolve a problem? after install, can't go to maintenance.pl because i get "Can't use an undefined value as an ARRAY reference at /usr/local/lib/perl/5.14.2/DBI.pm line 2058." 07:45 magnuse mysqld is not too bad, but i positively hate mysqlf ;-) 07:44 magnuse :-) 07:44 marcelr :) 07:44 marcelr ) 07:44 marcelr interesting typo, magnuse: no fan of mysql ? 07:41 magnuse myself-- 07:41 magnuse gah 07:41 magnuse mysqlf-- for breaking sample item types 07:41 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=10969 normal, P5 - low, ---, katrin.fischer, Signed Off , Fix sample itemtypes for translated installers 07:41 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=10965 major, P5 - low, ---, bgkriegel, Passed QA , Sample itemtypes can't load on new install 07:41 magnuse cait++ bgkriegel++ for bug 10965 and bug 10969 07:17 samueld ok, so, it can be changed 07:17 magnuse samueld: i would think the js was implemented as a feature, not by accident ;-) 07:17 dcook Time for me to run. Night/day all. 07:15 samueld is it a bug? 07:15 magnuse yeah, i was just thinking a couple days ago that forcing uppercase seemed a little heavy handed 07:14 magnuse ouch 07:13 samueld example, i write 'employee" and it is transformed in "EMPLOYEE", that's for why i've some problems to connect koha to the ldap 07:13 magnuse there is some js that transforms it to uppercase automagically, i think 07:12 rangi yes, its supposed to be samueld 07:10 samueld is it normal that when i create a categorycode in koha, it is written automatically in upper case instead of lower case? 07:10 j_wright pity mariadb doesn't work optimally unless you have ucontext 07:09 samueld hi magnuse 07:09 rangi hi samueld and reiveune 07:09 reiveune \o/ 07:08 magnuse bonjour samueld and reiveune 07:08 dcook hey samueld 07:08 samueld hi everybody 07:08 dcook rangi++ 07:08 dcook ^^ 07:08 rangi and that should be our priority 07:08 reiveune hello 07:08 rangi because for the library user, it gains them nothing 07:08 rangi if we get pg support as an aside, all well and good, but its not one of my priorities 07:07 rangi we are transitioning to maria 07:06 j_wright would be sad to be using DBIC and still be Mysql-bound 07:06 magnuse yay! :-) 07:06 rangi id agree with that 07:05 magnuse is what i'd say ;-) 07:05 rangi yep 07:05 magnuse stable master 07:05 magnuse dbic 07:05 magnuse plack 07:05 dcook Not that I disagree. I'm just curious :p 07:04 * dcook wonders what the massive priorities are 07:04 rangi but again, its not a massive priority 07:04 rangi yep, there arent many 07:01 j_wright well clearing out the mysqlisms would enable more than Pg support 07:00 rangi pg support is nice, but its not a massive priority 07:00 magnuse \o/ 07:00 rangi it currently works fine on mariadb too 07:00 dcook \o/ 07:00 rangi once thats pushed we can clean out the last of the mysqlisms, there are very few left 06:59 rangi you can deploy to postgresql, and it mostly works 06:59 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=8798 enhancement, P3, ---, elliott, In Discussion , Add the use of DBIx::Class 06:59 rangi with bug 8798 06:59 magnuse oops, listen to rangi, not me 06:58 rangi not entirely true 06:58 magnuse j_wright: there is none 06:58 j_wright what's the state of Pg support in Koha? 06:57 huginn magnuse: The current temperature in Marseille, France is 16.0°C (8:30 AM CEST on September 30, 2013). Conditions: Fog. Humidity: 94%. Dew Point: 15.0°C. Pressure: 29.65 in 1004 hPa (Steady). 06:57 magnuse @wunder marseille 06:57 huginn magnuse: The current temperature in Bodo, Norway is 8.0°C (8:50 AM CEST on September 30, 2013). Conditions: Mostly Cloudy. Humidity: 66%. Dew Point: 2.0°C. Windchill: 4.0°C. Pressure: 30.12 in 1020 hPa (Steady). 06:57 magnuse @wunder boo 06:57 magnuse tee hee http://www.nerdmeritbadges.com/ 06:48 Oak \o 06:48 wahanui hi, gaetan_B 06:48 gaetan_B hello 06:48 Oak Ahoy marcelr 06:47 magnuse hiya marcelr 06:46 dcook hey ya marcelr :) 06:46 marcelr hi magnuse dcook and oak 06:45 magnuse hehe 06:44 dcook kf++ 06:44 * dcook succumbs to peer pressure 06:42 magnuse kf++ 06:42 Oak kf++ 06:41 marcelr kf++ #qa 06:41 marcelr hi #koha 06:37 Oak hey dcook 06:37 Oak :) 06:37 * magnuse is too slow 06:37 magnuse Oak 06:34 dcook hey Oak 06:34 Oak kia ora #koha 06:34 Oak magnuse 06:31 magnuse hiya dcook 06:29 dcook hey ya magnuse 06:26 * magnuse waves 06:07 dcook I've had sneaky bad bibs get in and it's not pretty 06:07 dcook Especially if you're using the XSLTs 06:06 * dcook nods 06:06 mtompset I should, at some point in time, try to determine the exact record it is barfing on. 06:06 mtompset It does not reindex nicely otherwise. 06:06 mtompset It all indexes fine running this script. 06:05 mtompset Here's the thing. 06:05 mtompset It's not. 06:05 dcook Although you'll be glad when you've cleaned up the marcxml :) 06:05 dcook Doesn't sound like fun, mtompset 06:02 * mtompset grumbles about reindexing on 512MB of memory with no swap space and 53K+ records failing because of some malformed tag, and requiring using a script that inserts everything into the zebraqueue. 05:55 dcook If anyone wants a super easy sign off... 05:55 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=10974 normal, P5 - low, ---, dcook, Needs Signoff , OAI-PMH Resumption Tokens Do Not Handle Time 05:55 dcook bug 10974 05:53 dcook :/ 05:51 mtj ….and its been about since version 3.00.00 :p 05:50 mtj $version =~ s/(.*\..*)\.(.*)\.(.*)/$1$2$3/; 05:50 mtj # remove the 3 last . to have a Perl number 05:49 mtj theres a bug in the TransformToNum() sub, that stops 'local' versions from being evaluated correctly 05:42 dcook We track our db changes separately 05:41 dcook mtj: Maybe catalyst? 05:40 cait bbl 05:40 cait and i have to run and import patron data :) 05:40 cait we don't 05:39 mtj so, are any koha devs/companies actually using some system for 'local' version numbers for DB updates? 05:37 cait [off] all gone... i hope 05:35 cait [off] me still has limblime addresses in there for some people... email. 05:34 cait i really have to clean up my addressbook 05:34 cait args. 05:32 dcook Not always possible though 05:31 dcook Most of the time these days, I tell people that I'm happy to make changes to the database, but only after they've been vetted and pushed by the community first 05:31 cait we have been able to avoid it so far - but we are very cautious about local changes 05:31 * dcook tries to avoid local database changes these days 05:30 cait hm it depends 05:30 mtj sure, but also unavoidable cait 05:26 cait dcook: yay! 05:26 cait mtj: but local database updates are evil...;) 05:24 dcook huzzah. Changing the colons to slashes fixes the resumption token issue.. 05:20 dcook But the resumption token thing is still busted.. 05:20 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=10824 major, P5 - low, ---, dcook, Needs Signoff , OAI-PMH repository/server not handling time in "from" argument 05:20 dcook Maybe bug 10824 isn't relevant if we consider Koha to just handle a granularity of YYYY-MM-DD... 05:19 dcook Wait...no, it still wouldn't have worked... 05:19 dcook Now that I look at everything again...I think I finally understand... 05:13 dcook Maybe because no one else cares about oai :p 05:13 mtj hmm, there is a bug around the version number stuff - its not just me being stupid 05:13 * dcook feels like he's slowly becoming the oai guru... 05:12 dcook I guess I'll fix the resumption tokens... 05:09 cait it's ok ;) 05:08 * dcook is just unreasonably frustrated and didn't mean to direct his frustration your way 05:08 dcook I appreciate you telling me that :) 05:08 dcook Sorry 05:08 cait can only tell you the problems we ran into 05:07 dcook Showing deleted records wouldn't be too hard either 05:07 dcook Including item information would be easy 05:06 dcook The resumption tokens don't appear to work when using dates with times though 05:06 dcook Actually, no, I haven't seen that 05:05 dcook Yep 05:05 dcook These are fundamental core problems 05:05 cait and there was something about paging giving you an error when there were no more records 05:05 dcook Those are optional things though 05:05 cait but they were also not so happy about the missing deleted records 05:05 cait the big problem was that there was no option to include item information 05:05 cait they noted some problems about it 05:05 cait dcook: we had someone testing it for feeding their discovery interface 05:03 dcook Or maybe every one of my patches fixes one thing and breaks another.. 05:02 * dcook wonders if anyone uses Koha as an OAI-PMH server, because if they do...it's not working anywhere near correctly 04:59 wahanui hmmm... koha is a free software ils see http://koha-community.org for more info 04:59 dcook wtf Koha... 04:54 dcook argh.. 04:49 dcook I should probably listen to local radio but...I like the host and the show in Canada :/ 04:48 dcook hehe 04:47 wahanui okay, eythian. 04:47 eythian wahanui: new zealand is also in the future. We have jetpacks. 04:47 wahanui http://www.buzzfeed.com/daves4/hardest-parts-about-living-in-new-zealand 04:47 eythian wahanui: new zealand 04:46 wahanui http://www.buzzfeed.com/daves4/hardest-parts-about-living-in-new-zealand 04:46 eythian wahanui: new zealand 04:46 eythian so behind the times. 04:46 dcook Well, still Sunday in Vancouver..barely Monday in Toronto.. 04:45 * dcook would also be curious as to what you find, mtj 04:45 eythian not for too much longer though 04:45 dcook Oh noes...it's Monday... 04:45 eythian cool. let me know what you work out, too 04:44 mtj thanks for the info eythian - i'll poke about some more re: version numbers 04:43 * dcook also waves to cait 04:43 dcook 15 minutes until The Signal! 04:30 mtj morning cait :) 04:29 eythian oh, hi cait. 04:29 eythian I'm not sure if it's right, but it's how it should be done. 04:29 mtj fwiw - i did initially experiment with an extra '.' but got some weird error, so assumed i was doing it wrong :/ 04:26 eythian nope 04:26 mtj ahh, actually you didnt quite say that before… did you :) 04:25 mtj well, i thought the dots were simply ignored by koha (as you suggested?) 04:24 eythian you need a . for that :) 04:24 mtj yeah, well i thought i *was* adding another section on :) 04:22 eythian or increment the already existing minor number, but that's probably less ideal. 04:22 eythian I think you can just chuck another section on. 04:22 eythian ah no, that's internal 04:21 eythian so 3.12.05.002 04:21 eythian you have anohter section 04:21 eythian well hang on 04:21 eythian are you supposed to add another section or something? 04:20 mtj hmm, so whats a better string to give the 2nd 'local' update? 04:16 eythian the dots aren't decimal places, they're simply breaking up sections, and it's an integer in each section. 04:16 eythian in version number arithmetic 04:15 eythian mtj: 3.12.05.0001 == 3.12.05.001 04:15 cait eythian++ 04:15 mtj ...which is a gotcha if you are attempting to use 'local' version numbers, for your updates 04:14 mtj 3.12.05.001 04:14 mtj 3.12.05.0001 04:14 mtj 3.12.05.000 04:14 mtj so, for these 3 updates - the middle one gets skipped... 04:13 mtj the mentioned bug seems to be that CheckVersion() skips updatedb blocks with $VERSION values that are more than 3dp 04:06 eythian liberty is not the most sane software out there. 04:06 mtj wow, i cant blame you for not expecting that eythian 04:06 eythian mtj: in that case I guess CheckVersion should be fixed. 04:05 eythian I just found a situation my csvtomarc script can't handle. Embedded newlines in a title... 04:05 mtj . 04:05 mtj if ( CheckVersion($DBversion) ) { 04:04 mtj seems to work better than... 04:04 mtj if ( C4::Context->preference("Version") < TransformToNum($DBversion) ) { 04:04 mtj hmm, i've just spotted a subtle bug in updatedb.pl - for the version checking code… 03:41 bgkriegel bye 03:29 mtj cool, was my high version number, causing the problem :) 03:25 mtj yeah, there was indeed some info in there 03:24 wizzyrea but it might be in /var/log/dpkg.log 03:24 wizzyrea ah okies :) then it won't be in the apt logs. :) 03:24 mtj yeah, sorry - dpkg -i ./foo.deb 03:24 wizzyrea oh so did you install it with dpkg? 03:23 mtj (prolly testing a master build before) 03:23 mtj aah, i just noticed my 'version' was higher than my package 03:20 wizzyrea you installed the package using apt-get eh? 03:20 wizzyrea maybe it's different per apt config 03:20 wizzyrea hm it was in the one I looked at this morning 03:19 mtj perhaps that DB-upgrade info only gets sent to the screen, and not a file - i might be imagining it? 03:16 mtj hmm, thats blank for me 03:15 wizzyrea (at least it was this morning when I looked at it) 03:15 wizzyrea /var/log/apt/term.log 03:13 mtj a random pkg question… when installing a new koha package - where are the db updates logged too? 03:13 mtompset EVAR? 03:10 bgkriegel done ! :) 03:02 bgkriegel :) 03:01 bgkriegel hehe 03:01 wizzyrea your last comment EVAR?! NOOOOOO 03:01 bgkriegel i'm writing my last comment 03:01 bgkriegel in a minit 03:00 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=9611 enhancement, P5 - low, ---, srdjan, Needs Signoff , Changing the password hashing algorithm from MD5 to more secure Bcrypt 03:00 eythian bgkriegel: should bug 9611 have its status changed, or are you still working on it? 02:57 dcook Nor does Koha acting as a OAI server.. 02:56 dcook Koha does not... 02:56 dcook OAI-PMH repositories are supposed to use UTC timestamps 02:56 * dcook ponders 02:48 mtompset Hmmm.... 02:41 rangi i cant see why we would need a zip file 02:40 * dcook has no idea about Greybox 02:40 mtj ...they are unneeded yes? (and duplicated per lang and theme?) 02:39 mtj peeps - shall i log a bug to delete .the greybox/GreyBox_v5_5.zip file(s) in Koha? 02:11 dcook Although there are a few seconds between each request, me thinks 02:10 dcook I imagine 680 HTTP requests might upset any server... 02:10 dcook I am having an issue where I get a software error after harvesting 34,000 records from one Koha instance... 01:51 dcook I should actually revise that to say "from" and "until" parameters... 01:51 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=10824 major, P5 - low, ---, dcook, Needs Signoff , OAI-PMH repository/server not handling time in "from" argument 01:51 dcook bug 10824 is actually just awaiting sign off 01:50 dcook I might work on it someday, depending on how that union catalogue idea pans out 01:50 jcamins Doesn't seem to be. 01:49 dcook I guess it's not an active one though :p 01:49 huginn 04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=3206 enhancement, P5 - low, ---, gmcharlt, NEW , OAI repositry deleted record support 01:49 dcook bug 3206 01:47 dcook I think there's a bug for reporting deleted records at some point :p 01:42 jcamins dcook: nor does Koha report deleted records, making the use of OAI with Koha exceedingly problematic at best. 01:40 dcook brb 01:40 dcook At the moment, Koha doesn't 01:40 dcook And whether your repository honours that "from" date properly 01:40 dcook It just depends on how often you run your harvests and if you update your "from" date with each harvest 01:40 dcook The spec has selective harvesting built-in 01:40 dcook It depends on how you implement it 01:39 jcamins dcook: I may be mistaken, but my understanding of OAI is that it doesn't really allow for incremental updating. 01:37 dcook That might not fit your purposes though 01:37 rambutan well, hummm 01:37 dcook You could use OAI :p 01:34 jcamins It would allow for keeping a discovery layer synchronized without the trouble of setting up an actual push notification for changes. 01:34 jcamins Exceedingly useful. 01:34 wizzyrea yep, I just did it again, timestamp looks just fine in deletedbiblio 01:33 rambutan jcamins: that would be useful information? 01:32 jcamins It'd make me happy if I could run a report every day for all records touched in the last day. 01:32 wizzyrea cool 01:32 jcamins Yup. 01:31 wizzyrea that is what you meant right? 01:31 wizzyrea that was edit -> delete items, then edit -> delete record 01:31 jcamins Thanks. 01:31 jcamins Woohoo! 01:30 wizzyrea 2013-09-30 14:30:16 < record I just deleted timestamp from deletedbiblio 01:30 dcook I haven't tested though right now though, so I'd wait for liz on this one.. 01:29 * dcook has a bunch of badd deletedbiblio timestamps from 3.2 01:29 * dcook thinks that works these days but didn't back in the day 01:27 wizzyrea yep sec 01:27 jcamins The test plan is as follows: create biblio, note biblio.timestamp from MySQL client. Delete biblio using "Delete record" button in staff client. Check deletedbiblio.timestamp for that biblionumber. 01:26 jcamins I'm trying to confirm that deleting a biblio from the staff client updates the timestamp to the date deleted in deletedbiblio. 01:26 wizzyrea sure, sup 01:26 jcamins Hey, does anyone have Koha + MySQL on a test server to do a test for me real quick? 01:10 rambutan hi mtompset 00:54 mtompset Greetings, #koha. 00:50 wizzyrea that is completely silly 00:49 wizzyrea the fox relevant because well. Kids. 00:48 eythian I think that one's better anyway 00:48 wahanui OK, eythian. 00:48 eythian wahanui: stonehenge is <reply>http://youtu.be/mbyzgeee2mg 00:47 wizzyrea The fox is <reply> http://www.youtube.com/watch?v=jofNR_WkoCE 00:13 rangi :) 00:00 jcamins "Useful" implies "would be listened to." 00:00 jcamins Useful might be over-selling it.