Time  Nick          Message
23:59 pianohacker   rangi: cool, didn't know NZ libraries used overdrive
23:59 pianohacker   gmcharlt: do you have a second to discuss a dependency problem that affects 3.16?
23:58 rangi         pianohacker: http://ils.stdc.govt.nz/cgi-bin/koha/opac-overdrive-search.pl?q=harry%20potter
23:58 pianohacker   rangi: Yup. That reminds me, though
23:58 pianohacker   eythian: Ah, yeah, even just making ack ignore .po files like I did for ag reduced the difference a lot
23:57 mtj           a well worn road
23:57 rangi         pianohacker: you did the overdrive work eh?
23:56 mtj           jce, yeah - ive travelled your path a few times, myself :)
23:56 eythian       I had ack tuned to drop a large amount of unnecessary Koha directories
23:55 jce           mtj:  I do think this sounds like a plausible explanation for what is going on.
23:55 eythian       pianohacker: OK, when I make a .agignore from my .ackrc, it's a huge lot faster.
23:54 mtj           jce: you are making good progress :)
23:54 eythian       not on SSD
23:54 eythian       > time ack GetMarcFromKohaField
23:54 jce           May have to take a supper break before diving into this.  Thanks all for your help.
23:54 pianohacker   eythian: Just did, same result. What test are you running/are you on an SSD?
23:53 pianohacker   jce: heh. Just making sure :)
23:53 eythian       i.e. run each one twice ant least
23:52 pianohacker   (on my machine)
23:52 eythian       pianohacker: make sure you have a warm cache though
23:52 jce           pianohacker:  :O  shock and disbelief!
23:52 pianohacker   eythian: that's interesting. running a simple ag/ack MarcToKoha in my kohaclone is 0.47 vs 22.27 seconds
23:51 eythian       though perhaps there are differences in options
23:51 pianohacker   never a 100% guarantee
23:51 eythian       pianohacker: it's around half the speed of ack for me.
23:51 jce           mtj:  Ooh, I may check into that.
23:51 jce           Ok, so I should edit a couple records by changing their 090 tags  to 999, or copying them to 999 tags in marcxml, import those records, re-index Zebra, and see if I can search those records on Title, Author, etc.  If that works, I should fix the framework and Koha to MARC mapping, mass-edit the rest of the marcxml records, import them, re-index, and there's a 100% guarantee that everything will be happy in the world.  :)
23:51 mtj           http://git.kohaaloha.com/?p=misc/.git;a=blob;f=fix-090-to-999.pl
23:51 mtj           ahh, this script might do the 090->999 fixup jce
23:48 rangi         once we are sure that shifting the 090 to 999 fixes it, then we can worry about the mapping/frameworks :)
23:47 dcook         jce: Both cait and rangi have been doing this  much longer than me, so I'd probably listen to them :p
23:46 rangi         jce: everything else is moot if it doesnt
23:46 rangi         jce: id just try editing a couple of records, reindexing and see if it fixes the search
23:46 pianohacker   night cait
23:46 cait          :)
23:46 cait          night
23:46 pianohacker   dude is a bit performance crazy
23:46 pianohacker   eythian: see http://geoff.greer.fm/2012/09/03/profiling-ag-writing-my-own-scandir/ and http://geoff.greer.fm/2011/12/27/the-silver-searcher-better-than-ack/
23:45 eythian       pianohacker: ah, interesting
23:45 jce           dcook:  Ah, I was looking in the code.
23:45 cait          jce: i think before thinking furhter, just try changing the data for 2 records or so and see if you can search them - might save you more headache, doing it step by step
23:45 pianohacker   eythian: nope, even faster than ack, though not quite as full featured
23:45 eythian       ?
23:44 eythian       pianohacker: you mean ack
23:44 dcook         What you want to do is go to the page koha2marclinks.pl
23:44 pianohacker   dcook, eythian: It's a superfast recursive grep that automatically ignores certain directories
23:44 dcook         jce: a function. It's referred to in rebuild_zebra.pl
23:44 jce           dcook:  I don't find any reference to GetMarcFromKohaField in /usr/share/koha/intranet/cgi-bin/admin/koha2marclinks.pl  Is it a variable name?  A function?  A Perl module?
23:44 pianohacker   sorry, 80% of instances in Koha don't pass frameworkcode
23:43 * dcook       was also wondering this
23:43 eythian       what is 'ag'?
23:43 pianohacker   about 80% of the instances use it
23:42 pianohacker   well, from a very unscientific ag -i marctokoha | grep -v framework | wc -l
23:42 dcook         But yay for pianohacker looking at circulation! :D
23:42 * dcook       turns back to his Friday work before he starts thinking about other things..
23:42 * dcook       supposes it might depend on the case as well
23:41 pianohacker   always is in the library world, yeah...
23:41 cait          you never know with libraries
23:41 cait          there might be a use case i can't think of...
23:41 cait          i think it doesn't seem to make sense fo rme for different frameworks having different mapping right now
23:40 pianohacker   I've been looking at circulation code, so my head hasn't exactly been in the clean, recently architected part of Koha lately :P
23:40 pianohacker   (or at least an accepted one)
23:39 pianohacker   is there a best practice?
23:39 pianohacker   cait: that's a question I recently ran into, as some places use the marc mapping of the relevant framework and others just use the default
23:38 dcook         Yep. If those have 090 for the biblionumber/biblioitemnumber, new records and mods will get the number changed back to 090.
23:38 dcook         Not to be confused with nsync...
23:38 cait          you'd want to make sure that the mappings there are correct - they get applied to all frameworks
23:38 dcook         Whoa. Wahanui and I are in sync.
23:37 wahanui       /cgi-bin/koha/admin/koha2marclinks.pl is the place to check 1st
23:37 dcook         /cgi-bin/koha/admin/koha2marclinks.pl
23:37 cait          koha 2 marc mapping
23:37 cait          it's in administration
23:36 jce           dcook:  is GetMarcFromKohaField in system preferences?  I'm not finding it there.
23:35 dcook         Err "Koha to MARC Mapping"
23:34 rangi         (without fixing the frameworks, the problem will recreate itself over time)
23:34 dcook         Also want to update Marc => Koha mappings, as those get used quite a bit in Koha
23:34 dcook         That's what I meant. Bad communication on my part :p
23:34 rangi         need to fix the data and fraemworks to say 999
23:34 rangi         zebra looks for 999
23:34 rangi         the data and the frameworks say 090
23:33 rangi         thats the problem
23:33 rangi         yes
23:33 rangi         you will need to fix up your frameworks too jce
23:33 dcook         rangi: BIBLIONUMBER in: 090$c   BIBLIOITEMNUMBER in: 090$d is based on GetMarcFromKohaField
23:33 rangi         the problem is the are in 090
23:33 dcook         "GetMarcFromKohaField"
23:33 jce           rangi:  sounds like a good idea.  I have a working backup, the original server, and 2 virtual machines.
23:33 rangi         dcook: zebra is looking in 999
23:32 dcook         jce: I think you'll also need to change the place that is telling Zebra to look in 090$c and 090$d, I suspect
23:32 rangi         and see if it works, then do all of them (after a backup of course)
23:32 jce           dcook:  I think this is still leftover mess from years ago.
23:32 rangi         reindex
23:32 rangi         id try it on 2 or 3 records
23:32 jce           Would it work to do a global text replace in a marcxml dump, changing the 090 tag to 999?
23:31 dcook         So I'm guessing old Koha?
23:31 dcook         rangi: It looks like 090$9 is unimarc
23:31 cait          loc is not relevant here - this is koha specific
23:30 cait          if you can do that, i think it might work
23:30 cait          you should change the records
23:30 jce           So maybe I simply need to run a script to upgrade the schema?  [hoping for something easy like that]
23:29 jce           www.loc.gov/marc/bibliographic/bd09x.html says 090 is "Local Call Number [OBSOLETE, 1982]
23:28 rangi         hmm and old old koha too i think
23:28 rangi         ?
23:28 rangi         isnt 090 unimarc
23:27 jce           Ok all, here's another bit of information.  When I last tried to re-index Zebra, it reported the following:  BIBLIONUMBER in: 090$c   BIBLIOITEMNUMBER in: 090$d.  I take it that these should be in 999$c and 999$d?
23:20 dcook         hehe
23:20 pianohacker   dcook++
23:20 * pianohacker has a geeky trick
23:16 dcook         I don't think it would work without that prefix though as it would just interpret it as a kw search, me thinks
23:16 dcook         pianohacker: If you prefix your query with "pqf=" you can send straight PQF to Zebra
23:11 jce           rangi:  I'll need a little more context for your suggestion.  I've exported all the records in marcxml.  Are you just saying to check if I have 999 in the records.  I can already confirm I don't.
23:08 rangi         and have a look
23:08 mtj           me forgot that
23:08 rangi         is do a select marcxml from biblioitems limit 2;
23:08 mtj           aah, yeah - those arent automagically created :/
23:08 rangi         before worrying about frameworks
23:08 * eythian     wonders if there's a fixbiblionumbers type script around
23:08 rangi         the best thng to do
23:07 rangi         yep
23:07 cait          you'd need to add bilbionumbers i guess
23:07 mtj           or the other way around, too :)
23:07 rangi         mtj: that wont work if those records have no 999 on them
23:07 cait          the problem is, even after reloading the frameworks
23:07 mtj           me... i would load your bib, bibitems and items tables into a fresh koha, and test from there
23:06 cait          i agree
23:06 cait          yeah
23:06 cait          probably even load a standard framework
23:06 eythian       I'd be inclined to reload them from default
23:06 cait          i'd recommend looking at a standard default framework
23:05 eythian       that sounds like it would be very problematic.
23:05 cait          hmm
23:05 jce           So, I'm looking in Koha to MARC mapping, and I don't find any reference to 999.  I'll need to add it then?  Should this be under biblio, biblioitems, or items?
23:04 mtj           ..if ever?
23:03 jce           mtj:  That could be.
23:03 * tcohen      notices that hotel hasn't filtered port 22 :-D
23:03 mtj           jce: based on your info - you probably have not been using zebra for a while
23:02 tcohen        eythian: looking forward for it, thanks
23:01 tcohen        saturday actually
23:01 eythian       well, super-alpha tesing
23:01 eythian       some time soon I hope to squash patches appropriately and publish something for testing.
23:01 cait          tcohen: when will you be back? :)
23:01 mtj           ..its a common-ish bug for upgraded old kohas
23:00 tcohen        eythian: thanks for the update
23:00 mtj           Home › Administration › Koha to MARC mapping
23:00 mtj           jce, you are prolly missing 999 framework mappings - thats why your bibs have no 999 fields
22:59 eythian       that aspect of it needs a bit of work right now, but it's really just a bit of rearranging of files and such.
22:59 cait          jce: you ar elooking basically for 2 identical numbers
22:58 tcohen        awesome
22:58 eythian       tcohen: yep
22:58 jce           I see 700 tags that contain what look like author and title information.
22:58 tcohen        eythian: will there be something like K::SE::ES ?
22:57 jce           cait:  I'm looking through a MARC XML file I exported and it appears that you're right.  No 999 tags.  Could it be that we were never using Zebra even though we thought we were?  Seems more likely than that all the 999 tags got stripped out.  Is it possible that there was another tag that was being used instead.  I have vague thoughts about 7xx-something.
22:56 eythian       y
22:56 wahanui       i heard possible was everything, but I think it's probably easy to miss something here
22:56 eythian       possible
22:56 eythian       probably
22:55 pianohacker   wait, does that mean the existing code allows raw PQF for zebra?
22:55 eythian       e.g. title:foo AND subject:bar
22:55 eythian       this also means that you can type lucene stuff straight into the keyword box and it should work.
22:54 wahanui       somebody said interesting was sometimes good and sometimes bad
22:54 cait          interesting
22:54 eythian       If you're using the zebra-like code, it builds a lucene-type search string, and then turns that into an ES query.
22:54 cait          oh
22:54 eythian       It is
22:54 tcohen        so the search is not built byQueryParser
22:54 eythian       Yeah.
22:53 tcohen        instead of PQF
22:53 eythian       Basically you feed it a lucene search string.
22:53 eythian       It also has its own API, though it's a bit under-defined at the moment as I'm sorting out use cases.
22:53 tcohen        thanks eythian
22:53 tcohen        ok, so sticking to K::SE is a good way to do it
22:52 wahanui       okay, eythian.
22:52 eythian       it is also mostly using K::SE
22:52 eythian       I'm writing a compatibility layer so it can get requests from the existing stuff
22:52 tcohen        pianohacker: that's SolR
22:52 pianohacker   It's going through K::SearchEngine, no?
22:52 eythian       tcohen: there is, sorta.
22:52 tcohen        following that hypothetical API
22:51 tcohen        some people might like to work on making Zebra code a bit better
22:51 tcohen        (like there was for SolR)
22:51 tcohen        I was thinking, that if there was an API for implementing the ES code
22:51 tcohen        and also, that in the mid term, Zebra might even dissapear
22:51 eythian       s/refactored/taken out and shot/ but yes.
22:50 tcohen        we all know Zebra's code needs to be refactored
22:50 eythian       I didn't understand your question
22:50 tcohen        eythian: oh
22:50 tcohen        it is :-D
22:50 eythian       tcohen: ES API
22:50 mtj           tcohen:  Ecuador sounds nice :0)
22:50 tcohen        eythian: don't know what was about
22:49 cait          with the bilbionumber
22:49 cait          jce: the problem might be that your records are missing the 999 tag
22:49 cait          tcohen: have fun :)
22:49 jce           eythian:  the real problem is not how to start the server, but how to get it working correctly.
22:48 jce           I ran a bibliographic framework test a couple days ago and it said I had tabs 6 and 3 in use in tag 696, but when I went to try and fix it, I couldn't figure out how to get rid of one.
22:47 eythian       jce: though, I recommend 'sudo service koha-common start/stop/restart' instead
22:47 eythian       jce: koha-start-zebra
22:46 tcohen        #koha: i'm on a small trip to Ecuador, i'll try to push stuff to master soon, but not today
22:45 jce           cait:  It must be a data problem that has moved from the original server to the VM in the SQL.  A mapping problem would make sense.
22:45 mtj           yeah, a 999/mapping problem
22:44 cait          something about the biblionumber?
22:44 cait          i thnk it's a data problem
22:43 jce           mtj:  assuming koha-stop-zebra actually shut Zebra down, that's what's happening.  So how do I get Zebra running again?
22:42 mtj           jce: my hunch is you are not actually using zebra
22:40 mtj           $ ps -ef | grep zebra
22:40 mtj           jce:  double check
22:40 cait          can you still search when you flip the nozebra?
22:40 jce           mtj:  pianohacker:  I did 'koha-stop-zebra' and I can still search by "Library catalog" and get results.  Keyword searches still don't work.  Clearly Zebra is not working.
22:39 mtj           hmm, mapping (i think?)
22:37 jce           Is that a problem with mapping or frameworks?
22:37 cait          that's not good
22:37 cait          ah
22:37 jce           It exports 3295 biblios, which is plausible.  It gives a warning:  Record didn't contain match fields in (bib1,Local-number)
22:35 cait          also the -f i think
22:34 cait          -x -v -v for more output
22:34 cait          i think
22:34 cait          2 v
22:34 cait          one x to v
22:34 jce           cait:  The syntax on koha-rebuild-zebra:  -x -x -v -v, or -xx -vv?
22:34 mtj           nozebra = 'dont use'  should *not* give you any results, on 3.8.16
22:34 pianohacker   yeah, that's a good idea. Stop zebrasrv/koha-zebra if it's running, then try again
22:33 mtj           jce: if you stop zebra, and get search results... you have some problem :)
22:31 cait          does the rebuld give positive feedback? you could try adding the -x and a second -v
22:31 jce           mtj:  It was at one time.  It may be broken now.
22:31 eythian       hi
22:31 mtj           your koha has not been using zebra, perhap jce?
22:31 cait          it might be zebra is generally broken for some reason... and that only nozebra partially works
22:30 cait          does
22:30 cait          but that is really what the pref des
22:30 cait          they normally use zebra
22:30 jce           cait:  right.
22:30 jce           This suggests to me that my zebra is broken and "Library catalog" searches aren't using zebra.  Is this possible?
22:30 cait          hmmmm - but only library catalog searched, not the other options
22:29 jce           Ok, after I set it to "Use" I got no results in "Library catalog".  When I put it back to "Don't use" "Library catalog" searches work.
22:29 mtj           yeah, nozebra needs to be set to 'use' - thanks pianohacker, cait :)
22:24 pianohacker   jce: also restart the zebra server, just to be safe
22:24 pianohacker   mtj: since "Use" should set NoZebra to no or 0
22:23 jce           Ok, I'll set it back, restart apache, reindex (with -x) and see what that does.
22:23 jce           Mine was set to "Don't use."  But setting it to "Use" didn't immediately change anything.
22:23 pianohacker   jce: It should be set to "Use", from what we can tell
22:23 mtj           yes = 'dont use'
22:22 jce           mtj:  But in the web interface, the wording is "Use" or "Don't use" zebra search engine.
22:22 pianohacker   the wording is confusing, I hated describing the No* style prefs...
22:22 mtj           oops, i take that back ^^ :)
22:22 jce           Would I need to rebuild zebra's index to get the NoZebra preference to take effect?
22:21 mtj           according to searching.pref, nozebra  = yes *enables* zebra?!
22:21 pianohacker   jce: the only thing I can think of is to also pass -x to rebuild-zebra
22:21 mtj           jebus
22:21 pianohacker   ace
22:20 pianohacker   yup
22:20 mtj           the search code in opac-seach.pl refers to it
22:19 mtj           looks like its still live on 3.8.16
22:19 jce           It didn't seem to have an effect.
22:19 pianohacker   that's a good question.
22:18 jce           So to clarify, the NoZebra preference is already do-nothing cruft by 3.8, or do I need to make sure it is set right?
22:18 pianohacker   there are several .pref descriptions that I wrote that only partially removed confusion... things like gist, for instance
22:17 pianohacker   amen!
22:17 cait          end of confusion
22:17 cait          i am glad we finally removed that
22:17 cait          maybe you are right
22:17 jce           pianohacker:  -f -v {instance}
22:17 cait          pianohacker: i was wondering that
22:16 pianohacker   at least as of 8626a5bebb3f900d852a3d98c5f45d95cea5272d
22:16 pianohacker   and wait a second. From what I'm reading in the preferences file, shouldn't NoZebra be set to "use"? The .pref file accounts for the negation
22:16 jce           cait:  I see defaultSortField and defaultSortOrder and OPAC versions thereof.  Both are set to relevance and ascending.
22:14 pianohacker   hmm, okay. what flags are you passing to rebuild_zebra.pl?
22:14 jce           pianohacker:  No, it's MARC21 all the way.
22:13 pianohacker   jce: Do you have zebra set up for unimarc and are using MARC21, or vice versa?
22:12 jce           mtj:  No.  And again, I do get results when I don't specify Title, Author, etc.
22:12 pianohacker   jce: yeah, sorry about that. That and several other things were why it's since been eviscerated
22:12 cait          jce: hm the wording is irritating me too indeed
22:11 mtj           is your koha config pointing to another db?
22:11 jce           It's worded very poorly.  The pull-down says [Use|Don't use] the Zebra search engine.  But I'll set it back the way I found it.  It didn't appear to make any difference.
22:10 mtj           jce:  your search should not work at all, without zebra
22:09 pianohacker   jce: NoZebra should be set to don't use, it's old code that was removed not long after the version you were using
22:09 cait          in other words: you don't want to use NoZebra :)
22:09 cait          it#s double negated... it means use zebra then
22:09 cait          jce: Don't use NoZebra is correct
22:09 mtj           jce:  that is very very suspicious
22:09 jce           3.8
22:09 cait          jce: undo that
22:08 pianohacker   jce: which version of koha are you using?
22:07 jce           Ok, I searched 'search' in preferences and found that NoZebra was set to "Don't use".  Hmm.  I set it to 'Use' and saved Searching preferences.  Didn't seem to help immediately.  Do I need to restart apache to make Koha see that change?
22:05 cait          itshows your development :)
22:05 pianohacker   "What a dumbass idea. Who thought that was reasonable? ... Oh..."
22:04 rangi         :)
22:04 pianohacker   rangi: I was looking at the last attempt I did at ajaxcirc back in 2009 the other day like that...
22:03 rangi         http://kohadevreactions.tumblr.com/post/85406182518/coming-back-to-a-feature-you-wrote-months-ago
22:03 cait          that kid is cool
22:03 cait          aaw
22:02 pianohacker   http://devopsreactions.tumblr.com/post/88260308392/testing-my-own-code
21:59 pianohacker   hallo
21:59 mtj           jce:  no need for items
21:58 cait          look for search
21:58 cait          hm what is the default sort option set to in this instance? there are prefs for that
21:58 jce           mtj:  I'll have to add items against these bibs, correct?
21:58 cait          ah
21:58 mtj           ...then load those bibs into your problem koha, and confirm that you get different search results
21:57 mtj           ...then index, and confirm that the seaching behaves 'sanely'
21:57 jce           Ok, on the VM, the last error in /var/log/koha/BCMCKoha/opac-error.log is:  opac-search.pl:  Use of uninitialized value $sort_by[0] in join or string at /usr/share/koha/opac/cgi-bin/opac/opac-search.pl line 698.
21:56 mtj           load these bibs into your fresh instance -> www.loc.gov/catdir/cpso/RDAtest/extra_bib_marc.zip
21:55 mtj           jce:  yeah, a fresh instance
21:54 cait          but probably I'd take a look at the logs first
21:54 cait          both instances use the same koha code.. might give a clue?
21:54 cait          a fresh instance... and some records, see if that can be searched
21:53 jce           mtj:  by clean koha, you mean a fresh instance?
21:52 mtj           here is a dir of handy test records from LOC -> http://www.loc.gov/catdir/cpso/RDAtest/
21:51 mtj           prove that an ISBN search works on your clean koha, and work backwards
21:50 mtj           you really need to start testing from scratch -  against a clean koha, with some clean test records
21:49 cait          see if that gives you some kind of hint
21:49 cait          or run the zebrasrv in the foreground (can you say that in english?) and look at the queries
21:49 cait          youcould check the error logs
21:49 cait          jce: it would be my guess too... but really hard to tell :(
21:48 jce           So it must be something in the SQL.
21:48 jce           That's a fresh install of 3.8.
21:48 mtj           yeah, correct
21:47 jce           On the other hand, stale config files aren't likely to be plaguing the VM.
21:46 mtj           yeah, i use git for general sysadmin tasks too
21:46 jce           Guess that's another reason to learn how to use git.  :)
21:45 jce           Part of this long, sad saga is that I did upgrade from 3.8 to 3.14, but it didn't go well, so I put it back to 3.8 to fix it.  I'm now going to upgrade a VM first, then migrate that data to the production server.
21:44 mtj           ...then you can always do a nice before/after diff, to see what really changed
21:44 mtj           pro-tip: i use git to track /etc, before i do any upgrade
21:43 mtj           hmm, yeah, its possible
21:43 jce           mtj:  but it did start out life long before 3.8.
21:43 jce           mtj:  I did a "Complete Removal" of Koha in Synaptic, which I understand removes config files as well.  Is it possible that old zebra config could have survived?
21:42 mtj           ah ok... not too many zebra changes since 3.8 - thats good
21:41 jce           It's 3.8.  The package download file is 'koha-common_3.08.16.1-1_all.deb'.
21:41 mtj           jce, if its a really old koha, the zebra config files might be way out of date too
21:39 mtompset      There is that extra upgrade step if it is prior to 3.4 :)
21:39 jce           The VM is a fresh install that has the most recent SQL backup loaded.
21:39 mtompset      Is it 3.0.8 or 3.8?
21:39 cait          it's like i have seen something like this somewhere, but can't figure out where and what it was :(
21:38 jce           The system was originally set up in about 2007 or 2008 and has had several OS and Koha version upgrades.  I think that one problem came in when updating to 3.08.  Some of the MARC records have the infamous frey50 infection.  But I don't think that pertains to this problem.
21:37 mtj           jce: if you really get stuck - you usually need to set up a fresh/clean koha, and start comparing settings
21:36 mtj           Tools -> Koha to MARC mapping
21:36 jce           cait:  I doubt there was a change in the config, but I guess it could be.
21:35 cait          if it happened out of nowhere without updating
21:35 cait          so if the problem is in the configuration... might be a chance to find it there
21:35 cait          jce: action_logs should hav elogged changes to system preferneces
21:34 jce           mjt:  that's entirely possible.
21:34 jce           'm not exactly sure.  I'm an outside consultant who comes in to maintain a small church library's Koha server when problems arise.
21:34 mtj           jce, perhaps your mappings or frameworks are a bit wonky?
21:33 mtj           yeah, it is a bit weird :/
21:33 cait          you could check the action logs table for configration changes done in that time span
21:33 jce           cait:  I
21:33 cait          you didn't change any settings? when did it stop working?
21:33 cait          it's certainly weird...
21:33 jce           Well, I've tried that and not had good results.  I'm trying to fix 3.08 first, before upgrading.
21:33 cait          jce: running out of ideas sorry
21:32 mtj           ...and see if your search problems are fixed
21:31 mtj           jce: if i were you, i would load your 3.0.8 db into a new 3.14.x koha
21:30 mtj           morning all...
21:30 jce           Sorry, it reads differently between the 'real' server and the VM, but is consistent on each.
21:28 jce           Actually, as I look at it again, I don't think that's the problem.  On the 'real' server, the URL reads:  localhost/cgi-bin/koha/opac-search.pl?idx=au&q=searchterm
21:27 cait          really one is localhost and the other bcmckoha?
21:27 cait          hm that doesn't look quite right
21:23 jce           Hmm.  It does look different:  bcmckoha/cgi-bin/koha/opac-search.pl?idx=ti&q=searchterm  [bcmckoha would probably be an instance name, and may not be in /etc/hosts]
21:21 cait          how does the other url look like? for a title search or similar?
21:21 cait          so that works
21:20 jce           localhost/cgi-bin/koha/opac-search.pl?q=searchterm
21:19 jce           You mean a search under "Library catalog" in the pull-down?
21:18 cait          what does your search url for a non keyword search look like?
21:18 cait          hm
21:15 jce           Could it be a problem of items (holdings?) vs. biblio records?  Is that the difference between the "Library catalog" search and the others?
21:13 jce           SearchMyLibraryFirst is set to "Don't limit".
21:09 jce           OpacSuppression is set to "Don't hide".  It also says"  "Note that you must have the Suppress index set up in Zebra and at least one suppressed item, or your searches will be broken."  Does that pertain only if you have it set to 'hide'?
20:58 jce           Just getting the VM fired up again.
20:58 cait          turn the feature off ( don't hide records) if it's on
20:57 cait          did you check the setting?
20:57 jce           I know there are some bad biblio records that I need to fix, but I don't think this problem is related.
20:55 wahanui       well, opacsuppression is different.
20:55 cait          opacsuppression
20:55 cait          well opacsuppresson
20:55 cait          check administration > system preferences > opac suppression first
20:55 cait          yeah, looks like something in your configuration
20:55 jce           I have the same problem on the original server that I'm trying to fix, and on a server running in a VM that I loaded an SQL backup into.
20:54 jce           cait:  I'm not aware of having any of those options set.  Where are they found?
20:53 cait          are you using searchmylibraryfirst or a similar setting?
20:53 cait          jce: hmmm are you using opacsuppression?
20:51 jce           I'm working on a Koha 3.08 server that's having problems with OPAC.  When I do a search under "Library Catalog" I can get plausible results.  But if I select 'Author', 'Title', 'Subject', 'ISBN' or any of the other options from the pull-down list, I get "No results found!"  I've tried rebuilding Zebra many times.  Are the "Library Catalog" results not related to zebra?
20:14 nengard       got it
20:14 nengard       hehe
20:14 jcamins       nengard: yeah, I only thought of it because someone who learned from 3.2 tutorial videos and didn't need to mark anything lost since then (!) mentioned it.
20:14 rangi         morning
20:11 nengard       that happened a LONG time ago so I think I have that covered - not in a video but in a tutorial i wrote
20:10 jcamins       nengard: you used to do it in the additem screen but now you do it on the moredetails screen.
20:10 nengard       what what has changed about marking items lost?
20:10 nengard       jcamins - haven't had time to do videos in forever
19:23 huginn        cait: The operation succeeded.
19:23 cait          @later tell tcohen ping me when you got some time to talk?
18:58 cait          gmcharlt++
18:09 jcamins       nengard: are you still on the constant lookout for video suggestions? You may already have this one, but if not a video showing how to mark items lost on recent versions of Koha might be an idea.
17:04 reiveune      bye
16:55 gaetan_B      bye !
16:25 mtompset      Greetings, #koha.
16:04 oleonard      ...of course one would want to limit that query in other ways on a production system
16:04 pastebot      "oleonard" at 127.0.0.1 pasted "Querying deletedborrowers based on DELETE action in action_logs" (1 line) at http://paste.koha-community.org/36
16:01 barton        Do you mind posting that? I'd like to see what you did.
15:59 barton        hmm. tricky.
15:59 oleonard      From the reports interface. I ended up getting good results by using SUBSTRING
15:59 barton        (... reports interface => koha's reports page)
15:57 barton        oleonard: are you looking to run the report as a cron job or from the reports interface?
15:57 jcamins       I guess it's later than I realized.
15:56 gmcharlt      jcamins: indeed
15:56 huginn        jcamins: The operation succeeded.
15:56 jcamins       @later tell gmcharlt It occurs to me that it might be a good idea to have something in the code of conduct about cultural sensitivity, given the different traditions and cultures of members of the community.
15:31 * cait        waves
14:38 ashimema      oh dear.. how silly
14:29 paxed         or some other way to split the searchable text
14:29 paxed         use concat?
14:28 oleonard      Oh nice. You can't use a saved SQL report to query action_logs for 'DELETE' transactions because the query has the word DELETE in it :P
14:23 ashimema      I can't find the bug I was thinking of.. so must have wires crossed somewhere.. sorry.
14:21 ashimema      must be getting my wires crossed
14:20 ashimema      there was a bug about updating the timestamp in a dleeted table.. which I QA'd
14:20 ashimema      maybe it's deleted bibs I was tihnking of..
14:20 ashimema      but thats not the one I was thinking of..
14:20 huginn        04Bug 8926: enhancement, P5 - low, ---, gmcharlt, NEW , deletedborrowers should have a timestamp
14:20 ashimema      There's one.. http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=8926
14:20 oleonard      Oh, in action_logs it seems ashimema?
14:16 ashimema      the data is there.. somewhere.
14:16 ashimema      I'll did out the bug number for when it was all worked out.
14:16 ashimema      There was a bit of a debate around that oleonard..
13:54 oleonard      So the deletedborrowers table doesn't show a timestamp for when they were deleted, or am I missing it?
13:48 ashimema      hmmm. no cait today?
13:08 druthb        It was lovely!  Had a great weekend—busy busy, but good.
13:05 oleonard      I hope it was a nice one
13:05 oleonard      Hi druthb, and a belated happy birthday to you
13:00 druthb        o/
10:41 huginn        04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=11607 enhancement, P5 - low, ---, koha-bugs, NEW , items.content does not contain any formatting when HTML message is selected.
10:41 paxed         bug 11607
10:17 paxed         *grmbl*
10:16 sophie_m      paxed: cait it used to be an array formated, I don't know when it disappeared
10:11 cait          for the html, but that might be another notice... hm
10:11 cait          paxed: i might remember incorrectly, but i thought there was some code to bulid a table from it
10:10 paxed         also, not very html-firendly, that.
10:09 cait          there is a command line option for the fields it will output for items.content
10:08 cait          but yeah, it's tab separated, only overdues can currently do the nicer <item></item> formatting i think (haven't tested lately)
10:08 cait          to generate an item list
10:08 cait          it contains the fields you define
10:07 paxed         so, <<items.content>> in the notice will contain tab-separated lines of crap that's mostly useless to the patron?
10:06 paxed         'items.content' => $titles,
10:06 paxed         $titles .= join("\t",@item_info) . "\n";
10:06 * paxed       boggles at advance_notices.pl
09:21 Viktor        Hi cait :)
09:21 cait          hi Viktor :)
09:21 cait          it's a lot easier that way then trying to fix everything manually
09:20 cait          making sure the instance is marc21 is important so your data can be searched correctly /indexing configuration
09:20 cait          hope it all works out on second try :)
09:19 tony123       Great, thanks! I'll make the changes and reinstall.
09:18 cait          and then try reimporting
09:18 cait          then drop your instance or create a new one, make sure in the web installer, that you select marc21 as well
09:18 tony123       ok
09:18 cait          it should be marc21
09:18 cait          check what marcflavour is given there
09:18 cait          in the conf file for creating the instance
09:18 tony123       No problem. At least I have something to go on now :D
09:18 cait          that's cool then
09:17 tony123       I did
09:17 cait          it mght be best to redo the installatoin - did you install from packages?
09:17 cait          just switching the pref won't fix it all
09:17 cait          ok, now we got the problem
09:16 cait          marc21 can be unicode too :) we are using it with hebrew for example
09:16 tony123       oh, ok. wrong assumption. :)
09:16 cait          unimarc has nothing to do with unicode
09:16 cait          it's a misunderstanding that happens often
09:16 tony123       Ok, this gives me something to go on
09:16 tony123       We have Cyrillic titles, so we need to run unicode
09:16 cait          i am not sure which is the most common format in central asia
09:15 cait          so there are quite some important differences
09:15 cait          for example, the title in unimarc is 200 and the title in marc21 is in 245
09:15 tony123       ahhh
09:15 cait          my guess is that your database is unimarc but your data is marc21
09:15 cait          hm
09:15 tony123       Central Asia
09:14 cait          where are you located?
09:14 cait          oh
09:11 tony123       thanks. UNIMARC
09:08 cait          and search for marcfl
09:08 cait          administration > system preferences
09:08 tony123       Good question... where do I find it? I think I just selected the default flavour
09:06 cait          what's your marcflavour system preference set to?
09:06 cait          i am not sure reinstalling would make a difference
09:06 tony123       I did a manual cataluging. It went in fine.
09:05 cait          if you haven't made any configuration yet
09:05 cait          didyou try catalouging manually?
09:00 tony123       If you recommend a different distro, I can do that too
09:00 tony123       Do you recommend doing a clean install?
09:00 tony123       Same result: Citation field reads "null" Status reads "Staged". Match details and Records and both blank.
08:58 tony123       cait:: yes, the mrc file is largely unreadable.
08:57 tony123       I did create a record. I realized that I didn't create an Item type... I'll retry the import now.
08:56 tony123       Thanks BobB for your help! :)
08:52 BobB          diinner time, I'm off
08:43 cait          and when you open your file in a editor... it should be basically unreadable, then it's right :)
08:43 tony123       hang on, one sec. Let me check
08:43 cait          can you add a record there?
08:43 cait          tony123: when you go into cataloguing
08:42 BobB          hmm ...
08:42 tony123       If you think using a different distro will fix it, or reinstalling, I'm happy to do that
08:42 tony123       apt-get via the instructions on the website
08:42 BobB          by what method?  how did you install Koha?
08:40 tony123       Koha is installed on Ubuntu 12.04
08:39 tony123       MarcEdit's MARCValidator results: No errors were reported using the specified rules file.
08:39 BobB          :)
08:39 cait          and also what BobB says :)
08:38 cait          so there were no definitions for the fields and subfields in Koha - it's in the mandatory part of the web installer, but it's possible to uncheck it
08:38 BobB          Anyway, if it compiles ok, then it should mean your marc file is ok and your problem is a Koha one
08:38 cait          tony123: i had this happen when someone deleted the default framework once
08:38 BobB          are you sure you did that before?
08:38 cait          tony123: just for a test - can you catalog a reacord manually?
08:38 BobB          do that on the .mrk file, then recompile it to .mrc
08:38 wahanui       somebody said packages was at http://wiki.koha-community.org/wiki/Debian
08:38 cait          packages?
08:37 tony123       sure, thanks.
08:37 BobB          tony123, you could go back to MarcEdit and run the marc validator - that will show if your records are valid;
08:36 tony123       When clicked, the javascript error is: TypeError: $(...).html(...).dialog is not a function 	  title: _("MARC Preview")
08:35 tony123       the text is "null" with a hyperlink
08:35 tony123       the values in the the Citation field are all null. Status: Stages. Match details and Record and blank
08:34 cait          does the batch not show up at all?
08:34 cait          what do you mean by values?
08:34 cait          hm
08:34 tony123       the values in Staged MARC record management are all null
08:34 cait          on manage staged?
08:34 cait          but you say the title does not show up in the table below?
08:34 cait          that looks not bad
08:33 tony123       MARC staging results :      2 records in file     0 records not staged because of MARC error     2 records staged     Did not check for matches with existing records in catalog     0 item records found and staged     Manage staged records     Back
08:33 tony123       Here is the results log:
08:33 cait          your system is set up to use marc21? :)
08:33 tony123       no worries :)
08:33 cait          ok
08:33 cait          tony123: sorry to ask so many questions, but it's hard to tell - the message from thelogs probably is just a warning and not a problem
08:33 tony123       I used marcedit to convert it to the mrc file
08:32 cait          or with marcedit?
08:32 cait          tony123: did you convert it to an mrc file after using marcedit?
08:32 huginn        04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=11244 normal, P5 - low, ---, kyle, In Discussion , notices ignoring the dateformat preference
08:32 cait          bug 11244
08:32 tony123       I assume so... I used MarcEdit to build the file. I used the Z39.50 client to retrieve the ISBN data. It built file and decompiled file (back to the mrk file)
08:31 cait          i can find the bug for you
08:31 paxed         cait: ugh. and one of the most visible notices for patrons, at that. sucks.
08:30 cait          paxed: there is a bug for that - the problem was the the proposed solution slowed down the notice generation quite a bit, so it got stuck I think
08:30 cait          in iso format?
08:30 cait          are you importing a valid marc file?
08:30 cait          tony123: it#s hard to tell
08:28 paxed         does anyone else find advance_notices.pl formats the date-fields in wrong format?
08:11 tony123       Any thoughts? This is a clean install and trying to import he data for the first time
08:11 tony123       In Manage staged MARC records, the Citation field is null for each record. The staging results shows items as 0
08:10 tony123       Greetings. I'm trying to do bulk import (file generated from MarcEdit) an on Stage MARC for import, I'm getting this error: stage-marc-import.pl: Filehandle STDOUT reopened as FH only for input at /usr/lib/perl5/Template/Provider.pm
07:54 Tony          On Manage staged MARC records, the citations are null. Any thoughts?
07:53 Tony          This is a clean install and we are trying to populate the db for the first time
07:53 Tony          Greetings. I'm getting errors on staging marc for import: stage-marc-import.pl: Filehandle STDOUT reopened as FH only for input at /usr/lib/perl5/Template/Provider.pm
07:27 cait          hi fridolin
07:26 fridolin      hie all
07:22 gaetan_B      not so many here at biblibre though
07:22 cait          ah
07:22 gaetan_B      gaetan_B: hmmm, it's not really a holiday in France anymore but it used to be so a lot of people take a day off i guess
07:21 cait          everyone around on a holiday?
07:21 cait          hi gaetan_B
07:20 gaetan_B      hello
07:12 ashimema      brb..
07:05 * cait        offers cookies
07:02 * ashimema    wishes google chrome would 'just work' the way it used to on this Ubuntu box.
07:01 ashimema      moring cait, morning yohann
07:01 cait          *reads back*
07:01 cait          oh and hi Joubu
07:01 cait          hi yohann and ashimema
07:00 yohann        salut
06:54 ashimema      Morning Joubu
06:54 Joubu         Hi #koha
06:37 cait          oh nice :)
06:37 reiveune      here too, hi cait
06:37 cait          holiday here today :)
06:37 cait          hi reiveune
06:36 reiveune      hello
06:15 huginn        cait: The current temperature in Taegerwilen, Taegerwilen, Germany is 20.0°C (8:15 AM CEST on June 09, 2014). Conditions: Clear. Humidity: 83%. Dew Point: 17.0°C. Pressure: 30.12 in 1020 hPa (Rising).
06:15 cait          @wunder Konstanz
06:02 mtompset      Have a great day, #koha eythian wizzyrea mtj
06:02 huginn        mtompset: The operation succeeded.
06:02 mtompset      @later tell dcook bug 11592 is ready for testing now.
06:01 mtompset      dcook: 11592 is ready for testing. :)
06:01 mtompset      And with that horrible grammar moment, I think it is best to head to sleep.
05:46 paxed         *grmbl* fun, just noticed Yet Another Conversion Problem. the due dates were set with time 00:00:00, when they should've been 23:59:00
05:04 huginn        eythian: The operation succeeded.
05:04 eythian       @later tell tcohen how do you mean?
04:38 eythian       really it should be a biblio record I suppose
04:33 wizzyrea      nobody could find the bug with the old title.
03:36 huginn        04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=6874 enhancement, P3, ---, julian.maurice, Needs Signoff , Attach a file to a MARC record (Was: File upload in MARC)
03:36 mtompset      wizzyrea: That's another way of thinking about bug 6874. :)
03:21 mtj           cya jcamins
03:21 wahanui       If you feel like someone is looking through your window, it's OK, it's just me.
03:21 jcamins       Good night!
03:21 jcamins       :)
03:20 mtj           my brain hurts just thinking about that :0)
03:19 wizzyrea      ok yeah, that's not sensible.
03:19 wizzyrea      uhhhh
03:19 jcamins       They wanted two people... one to hold the laptop upside down, the other to type backwards.
03:19 jcamins       What's dumb is pointing it at a computer.
03:19 wizzyrea      I was like, well that seems sensible.
03:19 jcamins       No, that's a good idea.
03:19 wizzyrea      :)
03:18 wizzyrea      oh I misunderstood I thought you were upset that they were pointing it at a phone.
03:18 jcamins       wizzyrea: yes!
03:18 wizzyrea      wait they were pointing a document camera at a computer screen?
03:18 jcamins       Fortunately when a sizable portion of the participants said "oh well, I guess we won't present... when's dinner?" they decided to set up the projector hookup.
03:17 jcamins       eythian: exactly.
03:17 jcamins       I wouldn't have gone had I known.
03:17 eythian       it's ideal for a phone, because you can show ui. It seems dumb for a computer though.
03:17 jcamins       But there was no indication up-front that this was a mobile hackathon.
03:16 wizzyrea      the time I did this, there was no flicker, which may be why I'm a bit sympathetic to this
03:16 jcamins       wizzyrea: that's fine if you're presenting on a phone.
03:16 wizzyrea      yes, actually
03:16 eythian       in bad cases, you'd get flicker
03:16 wizzyrea      ok, then what is the proper way to put a phone interface on a projector?
03:16 jcamins       Ever pointed a digital video camera at a computer screen?
03:16 jcamins       Exactly.
03:15 jcamins       Which they brought in especially to project pictures of computer screens.
03:15 wizzyrea      but digital
03:15 jcamins       Yes.
03:15 wizzyrea      we are talking about a thing that is kind of like an overhead projector, right?
03:15 jcamins       And it wasn't a cost thing... I think they spent over $100k.
03:14 jcamins       They announced after everyone was there that they were using a document camera.
03:14 jcamins       The subject was a website.
03:13 jcamins       It was a hackathon. The entire point was that a bunch of people would be presenting interesting hacks they'd done.
03:13 wizzyrea      yeah context would be helpful ^.^
03:12 wizzyrea      more like "oh by the way we want you to show this giant room the mobile app"
03:12 jcamins       wizzyrea: oh, I missed some important context.
03:12 wizzyrea      jcamins: actually, if it's what I'm thinking of, and it was a live demo, I've actually done that. But I didn't have months to plan.
03:04 mtompset      Yes, that looks like it, rangi.
02:51 jcamins       wizzyrea: when you're the organizer and had months to plan, you don't get a bye that easily.
02:25 rangi         http://git.koha-community.org/gitweb/?p=koha-archive.git;a=heads all of those
02:24 rangi         if you havent run that in a long time, or git gc, you will have had all the old branches that were archived
02:17 eythian       It's probably for the best...
02:16 mtompset      Well, it worked, but I got a whole bunch of things "pruned".
01:59 eythian       try running git remote prune origin, though I don't know what it actually does.
01:59 eythian       yeah, it's a branch that became a directory with the same name
01:58 wahanui       Any ideas are welcome :)
01:58 mtompset      eythian: Any ideas?
01:58 pastebot      "mtompset" at 127.0.0.1 pasted "What does the suggested command mean?" (6 lines) at http://paste.koha-community.org/34
01:08 mtompset      Oh... it probably isn't a koha dependency. It's probably a package up a perl library issue.
01:06 * mtompset    grumbles about 2 minute boot time, because of wanting to be flexible.
01:05 mtompset      Just a second... I'll attempt something.
01:04 eythian       0
01:04 eythian       robin@zarathud:~/catalyst/koha$ grep -r Class::ISA * | wc -l
01:04 mtompset      (if I recall correctly)
01:04 mtompset      It was on an install attempt for me under Ubuntu 14.04
01:03 eythian       mtompset: Class::ISA doesn't seem to be a dependency.
01:01 huginn        mtompset: dcook was last seen in #koha 2 days, 18 hours, 8 minutes, and 3 seconds ago: <dcook> yo reiveune
01:01 mtompset      @seen dcook
01:00 mtompset      Greetings, #koha.
00:59 huginn        BobB: The current temperature in Sydney, New South Wales is 16.0°C (10:30 AM EST on June 09, 2014). Conditions: Scattered Clouds. Humidity: 59%. Dew Point: 8.0°C. Pressure: 30.36 in 1028 hPa (Steady).
00:59 BobB          @wunder Sydney, Australia