Time  Nick      Message
22:53 hdl       have a nice meal
22:52 chris     but now, lunch
22:52 chris     ok, just have to do a little more testing, but this should be merged before you wake up tomorrow
22:51 chris     ah cool
22:51 hdl       doing night house keeping on our servers.
22:51 hdl       ok for me.
22:50 chris     just fixed the copyright statements getting ready from the merge, its passing all the unit tests now
22:50 chris     anything you want to change there?
22:50 chris     http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=25295275972a605e71e532fb259cbdc7e13e3b8f
22:49 chris     but before you do
22:49 chris     go to sleep ;-)
22:48 hdl       still
22:48 chris     hdl: you still awake?
21:38 Brooke    a yep
21:38 ebegin    Where are you ? Virginia?
21:37 ebegin    Oh!  sorry, I though you were there.
21:37 Brooke    (But yes, course it is.)
21:37 Brooke    Selflessly, of course.
21:37 Brooke    I'll volunteer for the expedition.
21:37 Brooke    Em, I think that hypothesis requires in person exploration.
21:36 ebegin    Brooke, is it summer now in NZ?
21:34 ebegin    :)
21:34 Brooke    Silver lining :)
21:34 Brooke    but, your croissants are not crap.
21:34 ebegin    ... cold ...
21:33 munin     ebegin: The current temperature in Montreal, Quebec is -12.0�C (4:17 PM EST on December 15, 2010). Conditions: Light Snow. Humidity: N/A%. Windchill: -21.0�C. Pressure: (Rising).
21:33 ebegin    @wunder montreal quebec
21:33 ebegin    Hey!
21:33 * Brooke  waves at ebegin
21:15 munin     wizzyrea: Quote #71: "cait: hm it works now and I have no idea why :)" (added by chris at 07:47 PM, April 08, 2010)
21:15 wizzyrea  @quote random
21:15 munin     wizzyrea: Quote #107: "<kmkale> This is a food channel. Sometimes we discuss Koha too ;)" (added by jwagner at 02:49 PM, November 29, 2010)
21:15 wizzyrea  @quote random
21:15 munin     wizzyrea: Quote #10: "< pianohacker> You helped start an open source project; clearly your sense of what to avoid to make your life easier has been impaired for a while :)" (added by chris at 07:59 PM, June 23, 2009)
21:15 wizzyrea  @quote random
21:14 wizzyrea  woot!
21:14 munin     wizzyrea: The operation succeeded.  Quote #112 added.
21:14 wizzyrea  @quote add gmcharlt: only superlibrarians can be trusted with capes
21:14 Brooke    hmph. Oirish seems to have disapparated from Pootle.
21:13 munin     wizzyrea: Error: You must be registered to use this command. If you are already registered, you must either identify (using the identify command) or add a hostmask matching your current hostmask (using the "hostmask add" command).
21:13 wizzyrea  @quote add gmcharlt: only superlibrarians can be trusted with capes
21:13 wizzyrea  yay!
21:13 * Brooke  grins
21:13 gmcharlt  only superlibrarians can be trusted with capes
21:13 gmcharlt  apropos of nothing except wizzyrea's request
21:08 wizzyrea  I should mention that the change I proposed I'm not making this way, I was curious and clicked on it
21:07 wizzyrea  but I don't know if it's intentional
21:06 wizzyrea  I could fix this, easily
21:06 wizzyrea  yep, it's not those permissions, it's file system permissions
21:06 Brooke    are joo a superlibrarian?
21:05 wizzyrea  ohh.... bug or intentional: WARNING: You will not be able save, because your webserver cannot write to '/home/liz/kohaclone/koha-tmpl/intranet-tmpl/prog/en/modules/help/circ/view_holdsqueue.tmpl'. Contact your admin about help file permissions.
20:49 wizzyrea  yeap
20:49 nengard   those help files are usually copies of the manual
20:49 nengard   are we talking about a koha help file that's being edited? If so let me know so I can update the manual to match
20:32 wizzyrea  yup :)
20:32 chris     thanks
20:32 wizzyrea  will do
20:32 wizzyrea  right-o
20:32 chris     to tell them to
20:32 chris     probably should edit the help file for that page
20:31 wizzyrea  yep
20:31 chris     the keyword mapping?
20:31 wizzyrea  right
20:31 chris     for subtitle?
20:29 wizzyrea  chris: for 5143 do we need to update the database with those koha -> marc mappings or just document and trust people will add them?
20:28 druthb    Brooke++ #has her reasons, okay?
20:28 druthb    :-O
20:28 munin     druthb: Karma for "Brooke" has been increased 4 times and decreased 1 time for a total karma of 3.
20:28 druthb    @karma Brooke
20:22 hudsonbot * Galen Charlton: bug 5398: make additional pages in staff interface obey noItemTypeImages
20:22 hudsonbot * Nicole Engard: bug 4826 change 'add basket' to 'new basket'
20:22 hudsonbot Project Koha_Master build #221: STILL UNSTABLE in 22 min: http://hudson.koha-community.org/job/Koha_Master/221/
20:08 Brooke    Tēnā koutou
20:00 hudsonbot Starting build 221 for job Koha_Master (previous build: STILL UNSTABLE -- last SUCCESS #188 8 days 23 hr ago)
19:57 munin     New commit(s) kohagit: bug 5398: make additional pages in staff interface obey noItemTypeImages <http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=0560aa768cc1b128891259982a25ee2450470e79> / bug 4826 change 'add basket' to 'new basket' <http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=07834371f25ab207688e1534c39654049b890951>
19:54 briceSanc you're right !
19:54 chris     its quite low traffic, but thats where things like hte release schedule, and rfc etc are talked about
19:54 chris     devel is a good one to be on thats for sure
19:52 briceSanc i'm on request and patches, devel is missing
19:51 chris     ahh it went to koha-devel and koha-translate .. if you are working on either, those are 2 lists you should be on :)
19:51 briceSanc no i don't see this mail
19:50 chris     http://lists.koha-community.org/pipermail/koha-devel/2010-December/034871.html
19:49 chris     did you see the email?
19:49 briceSanc Do you have estimated the remaining time ?
19:49 chris     chris n sent a mail yesterday
19:48 briceSanc ok
19:48 chris     no, just template
19:48 briceSanc C4 freeze ?
19:48 chris     template freeze
19:48 briceSanc Does Koha is freeze in 3.2.2 ?
19:46 chris     yay
19:46 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=3523 normal, P5, ---, oleonard, ASSIGNED, Menu of existing lists limited to 10
19:46 * owen    is closing on a fix for Bug 3523
19:44 briceSanc hello all !
19:37 cait      colin++
19:36 owen      He managed to get so much done, yet we rarely see him around here. Is there a connection?
19:35 chris     colin++
19:35 chris     yup
19:34 owen      Colin++
19:18 chris     http://git.koha-community.org/gitweb/?p=koha.git;a=shortlog;h=refs/heads/new/enh/bug_5508
19:17 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=5508 enhancement, PATCH-Sent, ---, colin.campbell, NEW, Biblibre Serials work
19:17 chris     bug 5508
19:16 cait      clever :)
19:16 * chris   makes a bug for it, so he doesnt break his commits have a bugnumber rule
19:14 cait      I am working more with the old laptop (ubuntu) than with my newer one (vista) since kohacon
19:14 chris     ohh a pull request for the qa'd serials work, ill pull that down and do some testing on that myself then push it up
19:13 cait      lol
19:12 chris     friends dont let friends use proprietary software
19:12 cait      oh so many windows mails today
19:11 chris     owen, i basically said that
19:11 owen      wb chris and cait
19:10 cait      back
19:09 chris     back
19:08 drotsk    priceless
19:07 owen      minimum requirement system for windows: one Debian ISO </snark>
18:56 chris     Bb after coffee
18:56 chris     Ok my stop
18:55 cait      :)
18:54 chris     I'm sure she will
18:51 cait      hope she has fun
18:51 cait      :)
18:51 nengard   yeah i was in FL a week ago - panhandle - and everyone had scarfs and coats
18:51 chris     Kristina is going to her first cricket game
18:50 nengard   hehe
18:50 gmcharlt  consequently, everybody is dressing as if it were Anchorage
18:50 chris     I just hope it is fine on the 27th
18:50 gmcharlt  it has been below 0 C in Gainesville for the past week
18:46 nengard   I brought light layers :)
18:46 nengard   chris - yes you did - I don't blame you at all
18:46 chris     Jan/feb are the best months
18:43 cait      -thought
18:43 * cait    thought really liked the weather :)
18:42 chris     I told you layers sheesh :)
18:42 nengard   chris i was cold most days
18:39 munin     cait: The current temperature in Taegerwilen, Taegerwilen, Germany is -5.8�C (7:40 PM CET on December 15, 2010). Conditions: Light Snow. Humidity: 87%. Dew Point: -8.0�C. Windchill: -9.0�C. Pressure: 30.30 in 1026.0 hPa (Steady).
18:39 cait      @wunder Konstanz
18:39 wizzyrea  didn't keep us in
18:39 wizzyrea  it was windy and chillish but not awful
18:39 chris     December has been a lot more humid and rainy
18:38 cait      good morning chris
18:38 chris     I thought the weather was ok
18:37 nengard   looks like we should have had KohaCon in Dec instead :)
18:36 chris     Heh
18:36 brendan   ooh  chris squeaks out an early morning win
18:35 wizzyrea  <3
18:35 munin     chris: The current temperature in Wellington, New Zealand is 14.0�C (7:00 AM NZDT on December 16, 2010). Conditions: Partly Cloudy. Humidity: 88%. Dew Point: 12.0�C. Pressure: 30.09 in 1019 hPa (Steady).
18:35 chris     @wunder wellington nz
18:35 brendan   al
18:35 brendan   heya chris - wizzyrea et a;
18:35 wizzyrea  mornin'
18:35 chris     Morning from the bus
18:35 brendan   It seems that most of you are going to be jealous of my daily temps - so I'll make sure to check daily
18:35 wizzyrea  this last summer my car overheated on the hottest day of the year
18:35 nengard   or smoking engine
18:35 nengard   and the only thing that was hot was the burning engine
18:34 nengard   it was bitterly bitterly cold and windy
18:34 wizzyrea  AAA++
18:34 nengard   thanks - that was no fun
18:34 wizzyrea  nengard: glad you made it home safely from your car debacle yesterday
18:33 munin     nengard: The current temperature in CWOP # AR939, Penndel, Pennsylvania is -2.7�C (1:35 PM EST on December 15, 2010). Conditions: Mostly Cloudy. Humidity: 43%. Dew Point: -14.0�C. Windchill: -6.0�C. Pressure: 29.72 in 1006.3 hPa (Steady).
18:33 nengard   @wunder 19030
18:32 munin     wizzyrea: The current temperature in Channel 6 Downtown, Lawrence, Kansas is -1.1�C (12:34 PM CST on December 15, 2010). Conditions: Overcast. Humidity: 73%. Dew Point: -5.0�C. Windchill: -6.0�C. Pressure: 29.68 in 1005.0 hPa (Falling).
18:32 wizzyrea  @wunder lawrence, ks
18:32 munin     druthb: The current temperature in Woodley Gardens, Rockville, Maryland is -1.8�C (1:30 PM EST on December 15, 2010). Conditions: Clear. Humidity: 34%. Dew Point: -16.0�C. Windchill: -5.0�C. Pressure: 29.81 in 1009.4 hPa (Steady).
18:32 druthb    @wunder 20852
18:32 munin     brendan: The current temperature in Northwest Goleta, Goleta, California is 13.3�C (10:34 AM PST on December 15, 2010). Conditions: Overcast. Humidity: 87%. Dew Point: 11.0�C. Pressure: 29.96 in 1014.4 hPa (Falling).
18:32 brendan   @wunder 93117
18:06 wizzyrea  and I don't mean that as an indictment of how others do things
18:06 wizzyrea  but I understand why people would want the full name, since not everyone uses sane library codes
18:05 wizzyrea  I personally would like to have them both standard on the code :P
18:05 wizzyrea  owen: re 3262 I don't think the full name is being generated by the script for the homebranch
18:04 Brooke    Ka kite anō.
17:37 jcamins   So long, #koha
17:37 jcamins   Okay, time to get going.
17:31 Brooke    it's okay, though, soon Mr. Fairey will come and label it "obey".
17:31 cait      hm!
17:31 munin     cait: Karma for "koha" has been increased 17 times and decreased 0 times for a total karma of 17.
17:31 cait      @karma koha
17:31 cait      koha++
17:31 * Brooke  nods at jcamins
17:31 jcamins   (yes, it did)
17:29 owen      Sorry jwagner no idea
17:29 jcamins   Hey, didn't it used to be an Apple key?
17:29 jwagner   owen, I've been looking at the search code, but i can't see where it decides to use AND or OR.  Any pointers?
17:27 wizzyrea  lol ya
17:27 jcamins   Sneaky Command-W?
17:27 wizzyrea  oops
17:26 Brooke    because owen is not a whiner :D
17:26 * owen    hasn't complained enough about anything yet
17:25 jcamins   owen: has that worked for you yet?
17:25 cait      hehe
17:25 owen      Maybe Brooke thinks if she protests too much someone will break down and mail her some?
17:25 * Brooke  likes leveraging her pasta machine for labour savingses.
17:25 Brooke    bah
17:25 jcamins   Brooke: then you should know they're super easy
17:24 wizzyrea  these happen to have been made with sweet potatoes
17:24 Brooke    I've made them many times before
17:24 jcamins   ;)
17:24 jcamins   Errr... don't you?
17:24 jcamins   wizzyrea: wait, "why you don't eat gnocchi all the time"?
17:24 wizzyrea  eat gnocchi all the time
17:24 wizzyrea  and then you probably will
17:24 jcamins   See? wizzyrea can confirm
17:24 wizzyrea  like, so easy you'll wonder why you don't eat gnocchi all the time
17:24 wizzyrea  they are in fact super easy
17:23 jcamins   wizzyrea can confirm they're super easy.
17:23 jcamins   Somewhere in the scrollback is the recipe.
17:22 Brooke    lies!
17:22 jcamins   Brooke: they're super easy.
17:22 * Brooke  is jealous.
17:22 * jcamins just finished having gnocchi for lunch
17:22 jcamins   Mmmm.
17:22 wizzyrea  and the spud gobbled them up
17:21 wizzyrea  jcamins: we made the gnocchi last night -- Totally awesome, and super fast when you steam the potatoes
17:12 jwagner   Thanks -- finally deciphered my sketchy notes too
17:11 owen      ...for an example.
17:11 owen      jwagner: http://koha-community.org/koha-newsletter-volume-1issue-1-january-2010/#biblibre
17:06 jwagner   On 3693, maybe css would work better.  At the time I didn't know how to do it any other way.  Sysprefs might be easier for sites to implement, though.
17:05 jwagner   owen, the dev server where I did the shelving location work is gone.  Can someone remind me how to bring a branch down from a remote server?  I can see it on the git repo, how do I create a local branch with that code so I can work on it?
17:02 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=3693 enhancement, PATCH-Sent, ---, gmcharlt, NEW, Display options for buttons when holds triggered
17:02 owen      I wonder if Bug 3693 wouldn't be better implemented as a CSS customization
16:59 Brooke    owen, that is Munin's way of telling me to do "real work."
16:59 Brooke    Ko Omàmiwininiwak te iwi.
16:58 munin     owen: Error: "blew" is not a valid command.
16:58 owen      munin blew you far away Brooke
16:58 Brooke    Tēnā koutou
16:49 jwagner   http://koha.arcadia.edu/cgi-bin/koha/opac-search.pl
16:49 jwagner   owen, it's on Arcadia's OPAC (they sponsored):
16:48 cait      hi drotsk and wizzyrea
16:48 wizzyrea  lemme check, hi cait
16:48 drotsk    hi cait
16:48 cait      hi #koha
16:47 * owen    can see for himself
16:46 owen      wizzyrea: Is the shelving loc. search on your system?
16:45 jwagner   It's been so long since I worked on this, can't remember details.  I do remember a lot of trouble getting it to OR within the location set if you selected multiples.
16:44 owen      Searching by keyword + itemtype gives me an AND search
16:43 jwagner   No, that's one of the problems with combining searches, I think.
16:43 wizzyrea  i'm thinking of the other ones ;)
16:43 owen      Is that just me?
16:43 wizzyrea  for that one, it is
16:42 owen      I find myself expecting that if I search by itemtype and shelving location I expect it to do an AND search rather than OR
16:42 jwagner   wizzyrea, the jquery to hide it is in the bugzilla entry.
16:40 owen      It is in this case
16:40 wizzyrea  also work
16:40 wizzyrea  though I'm for adding that kind of stuff to the wiki
16:39 wizzyrea  probably the code to suppress something you've added should be in the bug report with the patch. But that's (a lot of extra) work
16:39 owen      I would be content to make it a default and wait for others to clamor for a system pref.
16:38 jwagner   Yes, on the advanced search screen.  Some sites don't use shelving locations.
16:38 owen      The presence of the shelving location search?
16:38 jwagner   You can hide it with a jquery -- I designed it that way -- but maybe a syspref would be better?
16:38 jwagner   And a question for you -- I didn't control this with a syspref partly because so many people were complaining about proliferating sysprefs at the time.  But do you think it should be?
16:37 owen      Hm, maybe not.
16:37 jwagner   No, I got results just by checking the box.  Dumb question -- did you search a location that you knew had items?
16:36 owen      Okay, now I'm getting results. I guess you can't get results if all you do is check the box but don't supply search terms
16:36 jwagner   Unless a site has modified record.abs (or maybe early sites didn't have that?) should work as is.
16:36 jwagner   melm 952$c      location
16:35 jwagner   Shouldn't have to -- shelving location is one of the delivered searches as far as I know.
16:35 owen      Do you need to modify the configuration of zebra?
16:35 jwagner   yes?
16:35 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=4329 enhancement, PATCH-Sent, ---, chris, NEW, OPAC search by shelving location option
16:34 owen      jwagner: I'm testing the patch for Bug 4329 and I have a question
16:34 jwagner   Maybe I can go back to New Zealand....
16:34 * jwagner definitely needs a vacation!
16:34 ibot      i heard jwagner was needing a vacation.
16:34 owen      jwagner?
16:27 slef      o/
16:11 Oak       \o
16:01 hdl       can't find an official instance of dev to test that.
15:57 hdl       can't find an instance.
15:57 hdl       mmm... I think that this is one of the information we got.
15:53 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=5157 normal, PATCH-Sent, ---, henridamien, ASSIGNED, borrowers top issuers filters problems
15:53 owen      Hi, I'm looking at Bug 5157 and willing to test the patch but I don't understand what the problem is
15:53 hdl       yes.
15:53 owen      hdl still around?
15:50 * munin   reloads and spins the chambers.
15:50 Brooke    @roulette
15:35 kf        :)
15:35 kf        hm or later
15:35 kf        see you all tomorrow
15:33 * Brooke  has faith that on Samhain eve, the Great Pumpkin will in fact appear.
15:29 owen      ...which might as well be addressed to "The Great Pumpkin" since no one is putting up money for it :)
15:28 owen      See also: http://wiki.koha-community.org/wiki/OPAC_Comments_RFC
15:28 kf        about bug 5500 :)
15:28 kf        nengard: I agree with you and owen
15:27 * Brooke  notes the Documentation Force is strong in this Channel.
15:26 gmcharlt  heh
15:26 Brooke    galen, behave.
15:24 gmcharlt  nengard: paint us a picture ;)
15:23 nengard   or imagining
15:23 nengard   good to know i'm not imaging things :)
15:23 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=5500 major, P5, ---, oleonard, NEW, shelf browse changing bib record
15:23 owen      nengard: I agree about Bug 5500--I don't think it used to work that way.
15:07 drotsk    o/
15:07 * Brooke  waves at drotsk
15:01 hdl       maybe we could serailize results in an array and use json or anything to display results.
15:00 hdl       jwagner: that would be good.
14:57 jwagner   i.e., a report of which barcodes were edited.
14:56 jwagner   hdl, and I've also had sites ask for output of the edit to go to a file or something more than just the screen display.
14:56 jwagner   We wanted to bring that capability in.
14:56 hdl       librarians need a validation of the selection before doing update.
14:56 jwagner   Yes, the ability to scan in barcodes at the batch edit screen is one of the things we liked about yours.
14:56 wizzyrea  it is far too easy to select items unintentionally, and there's no undo
14:56 hdl       Then it could be quite handy to reconcile both.
14:55 hdl       jwagner: if you would make the select interface a new page that would enter send  itemnumbers or barcodes to the tools in...
14:55 wizzyrea  Having seen both, NEKLS feels that the searching interface in the harley batch edit is too dangerous. We only train people to use the barcode entry portion of it.
14:53 slef      hdl: batch item edit
14:53 hdl       slef which ?
14:52 jwagner   There are parts of both that would be good to keep, question is how best to merge.  Which edit interface screen would be best? The one in our screenshot, or the BibLibre approach which looks like the standard item edit screen?
14:52 slef      owen: are we going with the liblime one or the parallel biblibre one?
14:51 jwagner   Was just talking about that one, as a matter of fact -- trying to figure out how to merge functionality with what's in 3.2 now
14:51 jwagner   It's still on my list to do
14:51 jwagner   Nothing at the moment -- we haven't had time to go back and pull it up or work on it.
14:50 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=3509 enhancement, PATCH-Sent, ---, gmcharlt, NEW, Batch item edit
14:50 owen      jwagner: You got an offer of help on Bug 3509. Did anything come of that?
14:49 jwagner   owen, was off in a meeting -- sorry, what were you asking about?
14:48 Brooke    http://www.grindtv.com/outdoor/blog/22880/diver+plunges+100+meters+unassisted+on+one+breath+to+set+world+record/
14:31 kf        ah :)
14:30 gmcharlt  kf: we do now :)
14:29 kf        gmcharlt: failed qa is a new status in the pull down - we don't have one for does not apply yet :)
14:28 kf        owen: perhaps set back to --?
14:25 owen      I have been making comments to that effect, just didn't know if we needed a specific patch status
14:24 gmcharlt  less strong of a statement, as a patch failing to apply is often due to a matter of timing
14:24 gmcharlt  owen: hmm - maybe mark it as 'patch does not apply'
14:18 owen      s/got/go
14:17 owen      If I got to test a patch for a bug marked "needs signoff" and the patch doesn't apply, should I mark it as "failed qa?"
14:12 thd       Unfortunately, I only found the official bug reports after scouring the documentation for configuration options which were not there to find.
14:09 thd       gmcharlt: I have determined that facet and truncation issues when using the ICU are confirmed bugs in the Index Data bugs database.
14:07 thd       gmcharlt: Some people had speculated that ICU problems with Zebra such as no facets were misconfiguration problems.
14:01 kf        ok
14:00 gmcharlt  gmcharlt at gmail.com
14:00 kf        gmcharlt: which email address?
13:57 kf        *sighs*
13:57 owen      s/Two/Six
13:57 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=4506 enhancement, PATCH-Sent, ---, katrin.fischer, NEW, Add support of record linking by record control number in $w
13:57 kf        bug 4506
13:57 kf        ... 4506
13:57 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=5406 enhancement, P5, ---, gmcharlt, NEW, Allow Koha to authenticate using LDAPS as well as LDAP
13:57 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=3509 enhancement, PATCH-Sent, ---, gmcharlt, NEW, Batch item edit
13:57 kf        ok, I will make a note - I have a sample file for my bug 5406 that I wanted to clean up - it has multiple languages too
13:57 owen      jwagner: Two months ago you got an offer of help on Bug 3509. Did anything come of that?
13:57 hdl       (they are committed in the solr branch iirc.)
13:57 gmcharlt  hdl: thanks
13:57 gmcharlt  kf: no problem, I can deal with the 880s
13:56 hdl       gmcharlt: I can send you some corean, some japanese hebrew unimarc records too
13:56 kf        and not sure that is indexed in standard
13:56 kf        they have hebrew in 880
13:56 kf        gmcharlt: I can do that - but there is a problem
13:54 gmcharlt  kf: would you mind emailing me a few of your Hebrew MARC records?
13:52 kf        but some search options in koha don't work with icu - I have to retest that sometime with 3.2
13:51 kf        we didn't see that with a few test records, but icu solved the problem and we were able to add rules ignore diacritics and such for seach
13:51 kf        term
13:51 kf        gmcharlt: standard indexing and hebrew records meant you got always all hebrew records back - with every hebre search
13:50 kf        the facets as is are a problem - not sure how much of it is zebra and what could be solved and what not
13:50 gmcharlt  thanks - that is the sort of specific information I'm looking for
13:50 kf        and using ICU means you loose some other features. but for our library it worked so far.
13:50 kf        gmcharlt: without icu it's not possible to search for Hebrew - at least we found no other way
13:42 thd       hdl: Can you help me distinguish better what cannot be done without using the ICU in Zebra?
13:41 thd       gmcharlt: Tokenisation without the ICU is based on single byte character encoding.
13:40 thd       gmcharlt: Without the problematic implementation of the ICU for Zebra, there is no tokenisation and some other support for more than circa 250 characters in Zebra.
13:38 owen      thd: Considering some of the errors in our MARC records, I'm not sure how well it worked...
13:38 thd       gmcharlt: I think that I still did not state it correctly.
13:38 thd       owen: However, I hoped that you would.
13:37 gmcharlt  thd: considering the hundreds of Koha installations that are successfully storing and searching MARC records using UTF-8, whatever applied at the time Zebra was adopted is a moot point
13:37 thd       owen: I doubted that you had the code.
13:37 owen      ...if anyone is still wondering.
13:37 owen      FWIW, NPL does not have a copy of a script to reconstruct 000 from other data in MARC
13:36 thd       gmcharlt: Zebra had no support for searching using Unicode when Koha adopted Zebra and we had not even realised.
13:33 hdl       icu is a friend of yaz, not really integrated.
13:32 hdl       gmcharlt: it lacks unicode support for search.
13:31 gmcharlt  Zebra lacks Unicode support?  that is demonstrably false
13:31 thd       hdl: We could have been ahead of the issue if we had been aware of lack of Unicode support at the outset.
13:30 thd       hdl: It is very unfortunate that we had not recognised the lack of Unicode support from the outset with Zebra.
13:29 kf        I would liked to see this happen in the community - the whole, those problem exist, we tried that that that...see bug ....
13:29 hdl       thd: it could have... But those problems were not encountered at that time...
13:28 kf        at the last meeting you said: biblibre made a choice
13:28 kf        hdl: yes, I see that you make an effort. I was talkign about the zebra problems you encountered.
13:28 thd       hdl: I think that some problems could have been addressed with the LibLime support contract with Index Data.
13:27 thd       hdl: I do not blame you.
13:27 hdl       we try to demonstrate.
13:26 hdl       thd: kf about communication, I donot think we did not communicate We are really willing to do things with the community. We organise meetings.
13:24 hdl       (maybe previous logs could give you the precise time)
13:23 hdl       (Paris time)
13:23 hdl       nengard: I think it was 9Pm ?
13:18 nengard   I could if needed but I'd probably need some instructions
13:17 nengard   slef I don't ...
13:17 nengard   someone who knows the time please update: http://wiki.koha-community.org/wiki/General_IRC_Meeting,_5_January_2010
13:16 slef      nengard: and do you have somewhere with a crontab that can send email? I can run it, but then you rely on me stopping/changing if/when you want.
13:15 slef      bbi60
13:15 nengard   creating a wiki page
13:14 slef      heh, topic
13:14 slef      Next general meeting on 5 January 2011
13:13 slef      next meeting time anyone off the top of their head?
13:13 nengard   yes please - nengard@gmail.com
13:12 slef      nengard: do I send to your on-list email address?
13:11 nengard   yes to anything that means helping nengard do her many jobs :)
13:11 nengard   yes to anything that means less work for me :)
13:11 nengard   reminders and scheduling
13:11 nengard   slef yes to it all
13:11 slef      nengard: yes to scheduling?  After I write the category idea, shall I base suggested reminders on your emails and send them to you for approval?
13:11 hdl       sorry nengard thanks for your action
13:10 nengard   slef hehe - scary stuff :) hehe
13:09 nengard   and hdl, actually I link to the RFC too in that newsletter so maybe we didn't need the two posts, but like I said I can't get enough content so I put them both in :)
13:09 slef      nengard: You really don't want to know how many emails we send co-op members... one member's email was down for a few months and I reconnected him yesterday and he had 2800 emails.  Our daily newsletter is a lot of that.
13:09 nengard   slef yes
13:08 hdl       or use the meetings logs.
13:08 hdl       nengard: the link could be done on the RFC around solr and the progress. iirc we posted that.
13:07 slef      want me to schedule them for you? ;-)
13:07 nengard   it's too much nagging for me :)
13:07 nengard   I'll do it monthly but I don't want to have to keep sending out 4 reminders to get articles
13:07 slef      I can appreciate the struggle for monthly, though.  It's more stuff that it's hard to get paid for.
13:07 nengard   well then maybe we need to put that to a vote too at the meeting
13:06 slef      nengard: today I got email about it from a librarian before I saw the newsletter myself.
13:06 slef      nengard: Shame, it seems like the newsletter is really popular, which I put down to your great work and persistent promotion.
13:05 nengard   well starting in 2011 it's a bi-monthly newsletter - it's too hard to get enough content from people for monthly
13:05 munin     slef: The operation succeeded.
13:05 slef      @later tell wizzyrea do we have feedwordpress on koha-community.org?
13:05 nengard   awesome
13:05 slef      yeah, I'll write something up
13:05 slef      then newsletter time, you'd tell it to update, see what comes in, delete crap, reorder what you want, then maybe just copy-paste the category index into the newsletter, with header/footer.
13:05 nengard   we might want to put it on the agenda for next meeting so others know the options/plan
13:04 nengard   that sounds fine with me
13:04 slef      I'm thinking set up feedwordpress to read category feeds like for-koha-newsletter (which I just imagined) into a koha-newsletter-proposals category on k-c.o and set it to manual update
13:03 nengard   oh cool - if we can do that then a-ok with me :)
13:03 slef      no, not all posts, just selected ones
13:03 nengard   if I post something Koha related on the ByWater site that is more ByWater related it shouldn't go to the k-c.org site
13:03 nengard   things get in that don't belong that way
13:03 nengard   and I don't like the idea of all posts from other Koha sites coming into the official site
13:02 nengard   yeah I don't have plugin access
13:02 slef      I'll look into it, if you're not against the idea.
13:02 nengard   not sure I have access to the plugins page ... off to check
13:02 nengard   I don't know ... wizzyrea would know that
13:02 slef      nengard: short blurbs are fine. Remind me: do we have feedwordpress installed?
13:01 slef      I think it's a bit odd clrh didn't link to the RFC or email thread where some claims in that article are queried, and other drawbacks are mentioned, but that's not an editing/compilation matter.
13:01 nengard   I'm up for any of these options
13:01 nengard   or I can start publishing entire articles in the newsletter instead of short blurbs (but I fear people won't read them)
13:00 nengard   I can say that if your info isn't on the official koha site then I can't include it
13:00 nengard   It would requires authors of posts in the newsletter to publish to k-c.org
12:59 slef      nengard: I'm not happy about the newsletter linking to biblibre.fr which requires one to pass an audio-visual ability test to comment.  Could we have newsletter articles hosted on k-c.o instead or is that a ton more work?
12:59 thd       hdl: I keep it a secret because not everyone would think the default configuration to be nice.
12:59 hdl       there is also iirc apache2-pazpar2 or masterkey for that.
12:58 hdl       nice...
12:58 thd       hdl: No I mean my own work in PHP which needs porting to Perl.
12:57 hdl       thd: about distributed solution... are you talking about the work that was in an unstable status (despite quite promising) from LibLime ?
12:57 thd       hdl: That is the worst problem identified thus far.
12:56 thd       hdl: I probably was not paying attention to Koha and I know that support companies not using the ICU may not have investigated well.
12:56 hdl       when you launch multiple concurrent queries... your server goes wild.
12:55 hdl       thd: realtime problems... zebrasrv using one proc for him alone.
12:55 thd       hdl: Do you mean a real time problem for CPU usage during queries or an indexing CPU problem.
12:54 hdl       kf: Taking decisions... about money or technical decisions... is a problem we raised... and that some told there was no problems.
12:53 hdl       We are trying to use zebra with no icu... in order to check whether it is icu.
12:53 hdl       For 300,000 records, it is not more a RAM problem than a CPU consumption.
12:50 thd       hdl: How much RAM do those libraries need?
12:50 hdl       thd: and we read that one library in India has 7M records.
12:50 thd       hdl: I think that we had another linguistic confusion.
12:50 hdl       thd: We have a library which has 800,000 records.
12:49 thd       nengard: Distributed Z39.50 searches with record matching.
12:49 nengard   thd whatcha got? :)
12:49 kf        it's nice to have the year in 000 for publication year seach in opac
12:49 thd       nengard: I have a better option for you.
12:49 kf        we normally use some default values if we can't make real fields
12:48 thd       nengard: kados had told me he spent two weeks working on the script.
12:48 nengard   okey dokey - at least we've confirmed some details
12:48 nengard   figures
12:48 thd       nengard: Perhaps NPL has a copy but it may be lost on some LibLime servers.
12:43 nengard   and is that script available in the wild?
12:41 thd       nengard: kados wrote a script for NPL to reconstruct at least 000 from other data in the records.
12:40 slef      nengard: oh, it'll also depend if MARC::Batch can read them without those!
12:40 thd       nengard: The original MARC Koha clobbered those fields as well.
12:39 thd       LibLime once paid for a support contract for Zebra with Index Data but such burdens should not fall on one company alone.
12:39 nengard   these are the ones that are missing
12:39 nengard   or one of the three
12:39 nengard   000, 007, 008
12:39 slef      nengard: erm. Ow. Which fields do you think are required?
12:38 nengard   this is so silly ... libraries used to delete the fixed fields to save space cause they were being charged too much, this is the second library I've had tell me this - so when you go to import their old records they don't work - stupid proprietary nonsense!
12:37 slef      nengard: looks like it depends on C4::AddBiblio which depends on the biblioitems columns being NOT NULL or some similar restriction. Checking that on a live server
12:36 thd       s/etc./etc. we would have formed a plan to fund fixing the bugs and/
12:34 kf        yes, I agree
12:34 thd       kf: I think that if there had been better communication by BibLibre with the rest of the Koha community about problems with Zebra ICU support etc. they would not have been in such a rush to add Solr/Lucene without preserving Zebra.
12:32 nengard   and thank you
12:32 nengard   yes please :)
12:32 slef      want me to check?
12:32 nengard   what about using the command line?
12:32 slef      nengard: might be able to import from Z39.50... I don't remember.
12:31 nengard   thanks for confirming
12:31 nengard   that's what I thought
12:31 slef      nengard: not from tools last I tried.  Would expect bulkmarcimport to barf too.
12:31 slef      sorry, didn't mean anything by it
12:31 nengard   data inmport question "Can a record import to Koha with no values in the three required fixed fields?"
12:31 slef      confusion
12:29 kf        but why kf #fail?
12:28 slef      I've been having a few problems with a courier.
12:27 slef      kf: http://identi.ca/mjray
12:27 kf        ?
12:27 slef      but still infinitely better than Initial Citylink
12:26 kf        :(
12:26 slef      kf #fail
12:26 druthb    :)
12:26 * kf      confused
12:26 kf        ah... he said it to me... right
12:26 kf        it was thd
12:26 kf        that was not my quote
12:26 kf        ah no
12:26 munin     slef: Error: "grrr" is not a valid command.
12:26 slef      munin: grrr
12:26 munin     druthb: The operation succeeded.  Quote #111 added.
12:26 druthb    @quote add <thd> kf: Many things can be funded with many people and small amounts.
12:26 munin     slef: Error: You must be registered to use this command. If you are already registered, you must either identify (using the identify command) or add a hostmask matching your current hostmask (using the "hostmask add" command).
12:26 slef      @quote add thd> Many things can be funded with many people and small amounts.
12:26 kf        huh
12:26 munin     kf: The operation succeeded.
12:26 kf        @hostmask add
12:25 nengard   hmm
12:25 munin     nengard: Error: You must be registered to use this command. If you are already registered, you must either identify (using the identify command) or add a hostmask matching your current hostmask (using the "hostmask add" command).
12:25 nengard   @quote add thd> Many things can be funded with many people and small amounts.
12:25 nengard   k
12:25 kf        want to try nengard?
12:24 kf        ... @quote add ...
12:24 kf        I have
12:24 thd       nengard: neither do I.
12:24 nengard   which i have no idea how to do
12:24 nengard   thd i like that quote - we need to add that to the quotes
12:23 kf        yeah, but it has to be discussed before development starts to make that work
12:23 thd       kf: Many things can be funded with many people and small amounts.
12:23 kf        and zebra is very fast in indexing
12:23 kf        thd: I am not against it, but as I stated previously - it should be an option - at least at first
12:22 kf        thd: perhaps that could have been funded by the community, lots of people, smaller amounts
12:22 thd       kf: Solr/Lucene is a good choice.  The implementation is merely problematic for adding the feature along side other features.
12:22 kf        and although I would love to have those solr features, I think we need to have an option for those real small libraries... and it will need time to prove a better solution.
12:20 kf        it's not like the community decided to go to solr - it's biblibre's decision. and the decision process is missing.
12:20 thd       kf: Money for a support contract with Index Data could fix those problems.
12:20 kf        d
12:20 kf        the problem I see is, that these problems are not really discusses
12:19 kf        the truncation and scan problems, yes. you have to live without some of the search features, but not sure nothing can be done about that.
12:19 thd       kf: I tried to qualify that statement correctly every time it was raised.
12:19 kf        and german umlauts, french diacritics
12:19 kf        with hebrew
12:19 kf        we are using it
12:19 kf        I can not confirm zebra + icu does not work at all
12:18 kf        I just read back
12:18 thd       kf: hdl and I were having multiple linguistic confusions during the meeting.
12:18 kf        the indexes get pretty big, but we have a lot of indexes.
12:17 kf        we once made a bigger database for a demo, I think 250.000?
12:17 kf        thd: I think the 8gb are disk space not ram
12:14 thd       slef:  Think of Biblios.net with millions of records indexed in Zebra.
12:13 thd       slef: If 300,000 records needs 8GB for Zebra that must not be a base number of records per GB or some services which use Zebra would be unaffordable to create.
12:11 thd       slef: It should not be in the sense in which I am asking.
12:08 slef      thd: might be commercially-sensitive info.
11:51 thd       hdl: Is 300,000 records the largest BibLibre library?
11:44 magnus    lunchtime!
11:37 magnus    good night reed
11:36 reed      g'nite all
11:35 BobB      Late here now.  I'm off.  Good night.
11:34 slef      once I free some disk space, I'll port it to HEAD
11:34 hdl       If you find it for HEAD then let me know.
11:34 slef      git@gitorious.org:koha/mainline.git for-3.0/bug_5394
11:33 hdl       Will try and make it in for 3.0.x
11:33 hdl       Sure ;)
11:33 slef      hdl: thanks but I can't find the bloody one for HEAD. Only for 3.0. Interested?
11:33 thd       hdl: Refactoring is not rewriting it is abstracting.
11:33 magnus    planning_and_sharing++
11:32 thd       hdl: I still think that we have a different conception of refactoring.
11:31 hdl       slef: which one ?
11:31 hdl       But it needs planning and sharing.
11:31 slef      Would someone review a patch branch for me, please?
11:31 hdl       This is why we try and make those refactorings part of new features.
11:31 thd       hdl: We need a better scheme for distributing costs where more than one library contributes to payments.
11:31 hdl       magnus: you are right.
11:30 magnus    i guess getting funding for writing tests is about as easy as getting funding for refactoring...
11:30 hdl       thx irmaB
11:30 irmaB     G'night all and thanks hdl for the good meeting.
11:29 hdl       But it costs... And Libraries have not been used to pay much on Koha developments. only the minimum
11:29 thd       hdl: I know that it has already been the case and that BibLibre have been great victims of the problem as well as perpetrators.
11:29 hdl       every single bug every single feature should have one.
11:28 hdl       thd: this has already been the case... And will be the case unless we have unit TESTS. for all.
11:28 thd       hdl: I have a concern that if we take the approach that refactoring will never be funded too much new development will be at the expense of old features.
11:28 hdl       But rather than letting ppl alone... working in small groups of interest could make better work in the end.
11:27 hdl       We will have work on Acquisitions and serials.
11:27 hdl       slef: we all contribute. There are many places to work on.
11:26 magnus    yeah, interesting meeting!
11:26 clrh      Thanks all! Happy to see more interractions today :)
11:26 munin     Log:            http://librarypolice.com/koha-meetings/2010/koha.2010-12-15-10.00.log.html
11:26 munin     Minutes (text): http://librarypolice.com/koha-meetings/2010/koha.2010-12-15-10.00.txt
11:26 munin     Minutes:        http://librarypolice.com/koha-meetings/2010/koha.2010-12-15-10.00.html
11:26 munin     Meeting ended Wed Dec 15 11:28:19 2010 UTC.  Information about MeetBot at http://wiki.debian.org/MeetBot . (v 0.1.4)
11:26 hdl       #endmeeting
11:24 slef      catalyst are funding template refactoring
11:24 tajoli    bye
11:23 hdl       on list please
11:23 hdl       And if you have other questions, or concerns or feed back on test instances, let us know.
11:23 tajoli    But if you need two charset, problems arise
11:23 hdl       I propose to stop the meeting now.
11:22 hdl       ok.
11:22 tajoli    And in this situation Zebra is good
11:22 hdl       ok.
11:22 tajoli    The english speaking countriies and contry like Italy with latin charset only
11:21 tajoli    Koha in fact is use on a simlgle charset enviroment
11:21 thd       tajoli: Yes that is correct.
11:21 slef      I think someone now is funding template refactoring (sorry I forget who).
11:21 tajoli    So every library with a complex charset (like Arab+latin) can't use koha in good way now
11:21 thd       tajoli: It works mostly but with important exceptions.
11:21 slef      hdl: that's not true. The co-op funded a lot of SQL-injection/placeholder cleanup way back when.
11:20 hdl       I explain ours.
11:20 tajoli    Is clear that Zebra+ICU doesn't work.
11:20 slef      hdl: that is your decision. Maybe this is easier for a private company. I just explain the difficulty of our membership organisation in the hope you will comprehend it.
11:20 hdl       libraries want features.
11:20 hdl       you donot want to redo things.
11:20 hdl       thd: noone would have ever funded refactoring.
11:20 tajoli    well, for me the main bonus of Solrs is to replace Zebra +ICU
11:19 thd       hdl: Do you not think you should have tried to obtain more funding for a development with such a large scope?
11:19 magnus    hdl: which is cool, but it seems there are lots of fun things to do with solr that may not be so important compared to getting solr into koha in the first place (which seems to imply making zebra and solr work as options next to each other)
11:18 hdl       And we did that.
11:18 hdl       slef: our work was only very little funded.
11:18 slef      hdl: it is difficult to persuade 90% of libraries that they should fund something to enable support for the biggest 10%, and I mentioned last meeting that I don't think our members will pay from the co-op's community donation fund. You see our difficulty here?
11:17 thd       clrh: I had looked at addbiblio.pl
11:17 hdl       and if all the RFCs cannot be integrated into 3.4 RM is fine with that.
11:16 hdl       But anyway, it is work on progress.
11:16 clrh      s/before/during
11:16 clrh      a record is constructed before indexing
11:16 tajoli    But for 3.6
11:16 hdl       tajoli: agreed..
11:16 hdl       magnus: we cannot. This is why we asker for help.
11:16 tajoli    Clearly is not a task for 3.4 (april 2010)
11:16 clrh      thd, you can look in Search::IndexRecord
11:15 hdl       And would have required too much time.... and would not have enable ppl wirh the flexibility we wanted to provide them.
11:15 thd       hdl: How do you return a complete record from Solr/Lucene?
11:15 magnus    hmmm... seems to me the main focus now should be on getting solr and zebra to both be options along side each other - otherwise it sounds like the solr solution will have a hard time becoming part of koha/replacing zebra...
11:14 hdl       It was proved not to be that efficient in indexing.
11:14 thd       hdl: I think that was merely a confusion between your description and a comment next to yours.
11:14 hdl       No.
11:14 hdl       yes... we investigated that. And think that we could build bridges.
11:14 tajoli    because you don't use solrmarc ?
11:13 hdl       tajoli: with solrmarc.
11:13 tajoli    Vufind has a MARC21 setup for Solrs
11:13 hdl       No.
11:12 thd       hdl: Did you state that you adapted an existing MARC 21 Solr/Lucene schema to UNIMARC?
11:12 irmaB     yes.
11:12 hdl       #action BibLibre make a solr instance for MARC21 and publicise that to Irma
11:11 hdl       #action irmaB build an adaptation of MARC21 for solr
11:11 reed      agree
11:11 thd       reed: Improvements in the sophistication of indexing may actually greatly increase RAM requirements for all options.
11:10 irmaB     I propose to look at the MARC21 implications with sorl - adapting the solr configuration for MARC21
11:10 tajoli    No
11:10 hdl       any other questions ?
11:10 hdl       josepedro: Please keep contributing...
11:09 reed      re: ram requirements --- my prediction is that solr schema  tuning is going to be a very long process and so and profiling done now is likely to go out of date fast
11:09 hdl       josepedro: Mod Perl Plack will be another meeting.
11:09 josepedro From our point of view, Koha has 3 big problems: 1- Facets. 2- Abstraction. 3- ModPerl. At present Zebra does not meet our expectations about facets, so we believe SOLR is the best solution and we would like to collaborate with BibLibre to develop this solution.
11:09 hdl       I am sure that Croswalking records in zebra is also ram demanding.
11:09 thd       hdl: how multiple?
11:08 hdl       thd: to be honest... it would require to do multiple tests.
11:08 hdl       #action provide a comparison of RAM requirements between zebra and solr
11:08 clrh      nop thd I ll try to provide it
11:07 thd       hdl: Do you have a comparison of RAM requirements in your Solr/Lucene test?
11:07 hdl       We will try and help him.
11:07 reed      hdl, right -- was going to say that I don't think you expect it should be a requirement for koha
11:07 hdl       and that is fine.
11:07 tajoli    yes, I confirm
11:07 slef      (I don't know OTTOMH, that is)
11:07 hdl       tajoli: propsed to help.
11:07 thd       hdl: wow, I have only tested very small record sets.
11:07 hdl       slef: reed we are willing to help but we cannot do that alone.
11:06 magnus    reed++ # sorry
11:06 magnus    redd++
11:06 slef      reed++
11:06 hdl       with koha3.2
11:06 hdl       thd: no. 300,000 with less that 8Gb is not viable option.
11:05 hdl       slef: if there is an abstraction layer you will have no problems.
11:05 thd       hdl: I think you missed a 0 in 3 million if you intended 3 million.
11:05 reed      snap
11:04 slef      hdl: I don't know how many records, but I suspect most koha libraries are smaller than that. Solr may be needed for the top 10% of libraries, but we cannot let 10% of libraries increase expense for the 90% unnecessarily, can we?
11:04 hdl       Even comments.
11:04 reed      slef, hdl - solr is just not going to be viable in small installations, it can't become a requirement for using koha
11:04 hdl       We are willing to recieve help.
11:04 hdl       We can help.
11:04 hdl       again, it is still work in progress.
11:04 hdl       #action josepedro send a mail with code reviews.
11:03 thd       hdl: Do you find database size limitations for Zebra?
11:03 slef      hdl: reportedly (see link I added to RFC), solr defaults to a 1Gb RAM usage.
11:02 hdl       And it would be quite nonesense to pretend that koha3.0 works in that context.
11:02 hdl       Those libraries surely donot have 300000 records.
11:02 hdl       slef: can you then be accurate in your demands ?
11:02 josepedro yes, no problem. But we have already sent you something about this.
11:02 * magnus  agrees with slef
11:01 thd       hdl: Solr/Lucene is not the API for library databases while Z39.50/SRU is.
11:01 slef      hdl: The co-op may be unusual in that we support almost as many self-hosted libraries as shared-hosted ones, but I would expect a lot of independent libraries to be worried by the increased resource demands of solr too.
11:00 thd       hdl: To Z39.50 servers?
11:00 hdl       solr has internal support for metasearch.
11:00 thd       hdl: What support do you envision for metasearch?
11:00 clrh      ok josepedro can you send us a mail with your points?
10:59 josepedro for example, the last i commented you
10:59 thd       hdl: One thing which would have worked on Koha using Pazpar2 which needs CCL is metasearch.
10:59 slef      yes, lots of our libraries have sub-gig servers
10:59 josepedro we consider it a great job but we think that there are several points that need to review deeply.
10:59 hdl       are you having servers with less than one gig ?
10:59 slef      but not use cases like "run in less than a gig" which are also a vital concern
10:59 hdl       #help please try and contribute yours
10:58 clrh      #link https://spreadsheets.google.com/pub?key=0AuZF5Y_c4pIxdEVzTjUtUGFoQnFpSkpfbTU5Ykc3b2c&hl=en&output=html
10:58 thd       hdl: Do not mistake that I do praise BibLibre's work.
10:58 hdl       We are gathering use cases of searches that worked in koha previously.
10:58 thd       hdl: Yet, not everything that worked before would work with the new system otherwise we would merely be busy praising your effort without these qualifications.
10:57 hdl       And then... use that abstraction to reintroduce options
10:56 hdl       Have any thing that worked before working with the new system.
10:56 hdl       thd: more abstract bricks so that it is more flexible.
10:56 thd       hdl:I have been typing furiously on that since last night in addition to other days.
10:55 hdl       We would like to have your feedback from the code review you did.
10:55 thd       hdl: safer ground?
10:55 josepedro we do not understand
10:55 hdl       But to build on safer ground.
10:55 magnus    good :-)
10:54 hdl       magnus: we donot want to break...
10:54 hdl       There is no other code to compare.
10:54 magnus    hdl: no one is denying that biblibre is doing good work here - it just seems odd to me that one of the biggest companies should introduce new "things" that break old "things" that a lot of people still want...
10:54 hdl       thd: it is not a question of comparison.
10:53 hdl       Can I add an action from xercode as of code review on what we did ?
10:53 thd       hdl: I do not question that much of the best work in Koha is work from BibLibre and Paul Poulain's business previously.
10:52 hdl       on the review.
10:52 hdl       josepedro: miguelxer we would appreciate your feedback
10:52 slef      #welovethenhs but it does mean I'm reluctant to waste public money by moving appointments
10:52 thd       hdl: I think I have understood your post about comparable code but stating that other work is inadequate should not be a basis for not attempting to develop a better model than other work.
10:51 slef      dentist, unavoidable
10:51 irmaB     but here :)
10:51 irmaB     late ...
10:51 irmaB     hi MJ
10:51 hdl       #help new ideas for plugins to add so that the indexing could be better.
10:50 * slef    = MJ Ray, worker-owner of software.coop
10:50 hdl       and many are using solr internally
10:49 irmaB     and proprietary LMS already offer (or say they do) facet searching and truncation etc.
10:49 tajoli    Zebra as is today
10:49 tajoli    Etc.
10:49 thd       hdl: I am trying to understand your last post.
10:48 hdl       #action work in pairs with CILEA for zebra as an option implementation
10:48 tajoli    No checks on data
10:48 tajoli    No indexes from administration
10:48 hdl       thd: show me any other code that works as much what we did. And I will be happy
10:47 tajoli    -- No improvment on facets
10:47 tajoli    -- No vanced search customization from administration
10:47 tajoli    -- No ICU
10:47 reed      tajoli, yeah, that sounds sensible
10:47 tajoli    clealy with Zebra no those options:
10:46 tajoli    But not in the same install, as an option to select in install
10:46 thd       hdl: Substituting one record indexing system for another is not refactoring as such.
10:46 hdl       We are willing to add advanced search customization from administration.
10:45 tajoli    I confirm, I will TRY to help to BibLibre to have Solr and Zebra as index tool in Koha.
10:45 thd       hdl: However, the particular implementation of refactoring C4::Search is BibLibre's work.
10:45 hdl       And would allow better end user experience.
10:45 hdl       magnus: we worked on that... And for solr integration... because it fixed many problems at once...
10:44 thd       ahhh
10:44 ibot      hdl: I forgot hdl
10:44 hdl       forget hdl
10:44 hdl       and is a point in 3.4
10:44 ibot      well, hdl is in France. France is in a galaxy far, far away.
10:44 thd       hdl?
10:44 hdl       magnus: C4::Search refactoring was not set by BibLibre.
10:43 thd       I should have s/on top/along side/
10:43 magnus    i do not like the sound of "others attempting to reintroduce Zebra, Pazpar2, etc. support on top of BibLibre's Solr/Lucene work" - sure biblibre has put a lot into this, but why should "others" have to re-implement something that is working (although not perfectly) today?
10:43 thd       hdl: Do you understand the problem that anyone starting from your work on C4:Search to add other non-Solr/Lucene options would need to rewrite all search calls to keep your Solr/Lucene work?
10:42 hdl       ?
10:42 hdl       tajoli: said he would help us in making solr an option... any other persons  ,
10:41 tajoli    With Solr as option I don't suggest to use Zebra with ICU
10:41 reed      (I might be exagerating the case a little)
10:41 hdl       We chose the best bricks for that.
10:41 reed      solr ram reqs will vary depending on updates and number of indexes and searches and catalog size and it'll be a few years before we have some stable config advice
10:41 hdl       Refactoring the C4::Search was a priority for us.
10:41 hdl       reed: but we cannot do that alone.
10:41 hdl       reed: I always said that we would like to build that.
10:40 tajoli    Ok, thank you.
10:40 hdl       that you needed to set zebra on a different machine.
10:40 hdl       If you read some logs about performance improvements, mason pointed
10:40 tajoli    I don't see it.
10:40 thd       tajoli: Zebra certainly has ICU but it does not work for scan queries for facets nor truncation other than right truncation.
10:40 hdl       absolutely.
10:39 tajoli    are you sure ?
10:39 hdl       bug CPU consumption is for zebra.
10:39 tajoli    big CPU consumption with solr ?
10:39 hdl       big CPU consumption
10:39 tajoli    We can't ask to improve RAM requests
10:39 clrh      I have to see how much ram is needed with solr, didn't test anymore
10:39 reed      hdl, good to hear -- but still is a thing that sounds like it needs additional attention
10:38 tajoli    For use this is the kay point
10:38 tajoli    Less RAM to use
10:38 tajoli    but
10:38 tajoli    -- bad facets
10:38 tajoli    -- no ICU
10:38 tajoli    -- diffcult indexes setup
10:37 magnus    hdl: we all have limited resources
10:37 tajoli    But with pointing that Zebra continues to have the problems:
10:37 hdl       Well reed it is not much separate. I think it could be built on top of what we did.
10:36 tajoli    sorry Biblibre
10:36 reed      there's a separate project here and I think that's 'make koha support various search back ends'
10:36 tajoli    Well, what I want to say is that CILEA can TRY to help biblibre to develop an abstract search call interface
10:36 hdl       We are willing to share ideas and development.
10:36 thd       Would BibLibre not at least consider abstracting the search calls so that others attempting to reintroduce Zebra, Pazpar2, etc. support on top of BibLibre's Solr/Lucene work would not entail rewriting BibLibre Solr/Lucene work for better abstraction?
10:35 hdl       our main problem is that we have limited ressources.
10:35 hdl       magnus: yes. Sure.
10:35 hdl       tajoli: koha 3.0 was claimed to support full utf8 search
10:34 thd       tajoli: Zebra is fixable but with support money.
10:34 tajoli    And Zebra + ICU doesn't work
10:34 hdl       This is how we ... and the whole community came into that problem.
10:34 magnus    so what is the real question here? no one seems opposed to solr as such, but there are some good reasons for keeping zebra around too. as long as solr is introduced as an option along side zebra everyone is happy, right?
10:34 thd       tajoli: As hdl states big libraries need full Unicode support.
10:34 hdl       in Hebrew, arabic... and Georgian..
10:34 hdl       even small have some special books
10:33 thd       tajoli: That is exactly how BibLibre came to their problem.
10:33 hdl       whatever quite big library you are.
10:33 tajoli    But I undestand that Zebra + ICU is mandatory  with more that one charset (like in France).
10:32 thd       tajoli: Zebra has a few problems but we should be able to have both Zebra and Solr/Lucene together.
10:32 tajoli    and realtime indexing work (with a daily reboot)
10:32 josepedro when you get the solr records,  you search them in the database instead of printing them directly
10:31 tajoli    For me setup Zebra is not a problem
10:31 tajoli    Well, as I have undestood the mains problems are Zebra + ICU and facets.
10:31 thd       hdl: Where did you find a Solr/Lucene configuration for MARC 21?
10:31 hdl       We are using records. getting them from koha and processing information to send to solr.
10:30 hdl       josepedro: I donot understand.
10:29 hdl       Plugin system we implemented for solr (that could be also used for zebra) is quite handy
10:29 josepedro without looking the database??
10:29 josepedro have you seen something about printing solr records directly??
10:28 hdl       we based our work on UNIMARC)
10:28 hdl       and adapting the solr configuration for MARC21 (
10:28 hdl       josepedro: it looks that we have a nice solution... With true facets. needs some more work on thorough configuration of indexes.
10:27 josepedro yes, we would like to collaborate with you
10:26 josepedro anyway, our main aim is facets solution
10:26 hdl       Do you have a plan or could devote people to work with us ?
10:25 josepedro we think that the implementation is not going to be very difficult with time enough.
10:25 hdl       and code reviews ?
10:24 hdl       josepedro: what have you figured out  from your audits ?
10:24 josepedro libs for dspace
10:24 hdl       josepedro: what have you found ?
10:24 josepedro we are reviewing the sru/srw libs, too
10:23 thd       magnus: SimpleServer will map CQL to PQF for SRU.
10:23 clrh      #link http://www.nntp.perl.org/group/perl.perl4lib/2010/12/msg2836.html about another idea for improving => multi-threaded call
10:22 hdl       - forking and sending // batches to index to solr.
10:22 magnus    ok
10:22 hdl       magnus: I guess and hope that SimpleServer will also cope with SRU
10:22 josepedro since Biblibre started with solr, we started to review their code comparing with Vufind
10:21 magnus    and SRU?
10:21 hdl       (posted from erik hatcher...)
10:21 clrh      #link http://lucene.472066.n3.nabble.com/Getting-started-with-DIH-td504691.html
10:21 hdl       - DIH Marc
10:20 clrh      #link http://www.indexdata.com/yaz/doc/tools.html
10:20 hdl       But we are working on two ideas.
10:20 hdl       indexing speed is still quite slow compared to zebra...
10:20 josepedro yes
10:19 hdl       #action improving the indexing time.
10:19 irmaB     josepedro was that with Koha 3?
10:19 josepedro we started with pazpar but we thought that is not the best solution
10:19 hdl       #action Z3950 support on top of solr
10:18 hdl       clrh: mentioned that we worked on the Z3950 grammar to get the whole of it (at least what is presented on the Indexdata website which was the only ressource we got that from)
10:18 hdl       Z3950 support on top of solr.
10:17 hdl       What we have problems and are willing to do ...
10:16 hdl       fredericd: miguelxer and josepedro can you preset your selves for the records ?
10:16 munin     04Bug 2048: blocker, PATCH-Sent, ---, gmcharlt, CLOSED FIXED, Kohazebraqueue daemon MAJOR issue
10:16 thd       tajoli: #link http://bugzilla.indexdata.dk/show_bug.cgi?id=2048 which is a bug for no facets with the ICU in Zebra which Index Data will only fix with a support contract.
10:16 josepedro that's true. we have proved it.
10:16 hdl       and all the setups are a little bit different.
10:15 hdl       And CCL is not a standard... It needs configuration
10:15 fredericd it's not a solution
10:15 fredericd tajoli: if you want accurate facets with Pazpar2, you need to send to it the whole resultset
10:15 thd       tajoli: Pazpar2 requires CCL support which BibLibre have removed in their implementation.
10:15 hdl       pazpar2 could have been chosen... We chose solr because the community is much more active.
10:14 tajoli    Ok is a dead road ?
10:14 tajoli    why not insert Pazpar2 as mandatory and use its facets ?
10:13 tajoli    One of the problem that you report is about Facets
10:13 hdl       listening.
10:13 tajoli    I have a suggestion about Zebra in Koha
10:13 hdl       mega bytes.
10:12 thd       hdl: 17 mo?
10:12 hdl       or catalogue.solr.biblibre.com
10:12 irmaB     wow that is impressive
10:12 hdl       as you can see and test on solr.biblibre.com
10:12 hdl       - display items information is now OK.
10:12 hdl       - Configuration of indexes via database is now ok.
10:11 hdl       - advanced search is now working pretty well.
10:11 hdl       What we achieved
10:11 hdl       (well not all of those are interestiing... But we plan to take out the most relevant ones)
10:10 hdl       magnus: we are doing some work on that... And will use 17 Mo of real use cases z3950 RPN queries to validate.
10:09 clrh      we have to match both together and translate in solr requests
10:09 hdl       #help on building Data::SearchEngine::ZOOM and Data::SearchEngine::Query::PQF via ZOOM wrappers
10:09 clrh      #link https://github.com/eiro/rg-z3950-rpn/ wip about rpn grammar and z3950 server development
10:08 hdl       And we would be grateful if the community would help us inahcieving that.
10:08 thd       magnus: #link http://git.biblibre.com/?p=koha;a=blob;f=misc/z3950.pl;hb=refs/heads/wip/solr misc/z3950.pl shows BibLibre work on a Simple2ZOOM gateway.
10:08 hdl       That would be excellent not only for Koha... But also for Data::Search::Engine.
10:07 hdl       And I think that we could wrap ZOOM in them
10:07 hdl       But bricks that we used are flexible.
10:07 hdl       Because of time and ressources.
10:07 hdl       our work now is not an either or option...
10:06 hdl       guetting back then.
10:06 magnus    i guess my main concern is support for Koha acting as Z39.50 and SRU server - what is the status on that and Solr?
10:06 thd       Does BibLibre intend to do work towards refactoring existing Koha record indexing and retrieval for use alongside Solr/Lucene and not as an either/or option?
10:06 tajoli    Zeno Tajoli, CILEA
10:05 hdl       #topic what is done
10:05 hdl       three
10:05 hdl       two
10:05 clrh      #link http://www.biblibre.com/en/blog/entry/solr-developments-for-koha
10:05 hdl       no questions one
10:05 hdl       Since there are no questions then we will skip to next topic : what we did and are up to.
10:05 clrh      #link http://wiki.koha-community.org/wiki/Switch_to_Solr_RFC
10:04 clrh      #link http://librarypolice.com/koha-meetings/2010/koha.2010-12-07-20.00.log.html
10:03 hdl       then ask.
10:03 hdl       If you have any questions or doubt on what we said previously and posted on list,
10:02 hdl       I think that this topic has been long advocated...
10:02 hdl       #topic Why taking on solr
10:01 miguelxer buenos dias a todos!!, ja
10:01 hdl       it is presentation time
10:01 ibot      bonjour, miguelxer
10:01 miguelxer hello
10:01 hdl       hi miguelxer
10:00 hdl       any other persons ?
10:00 magnus    Magnus Enger, Libriotech, Norway
09:59 Joubu     Jonathan Druart, Biblibre
09:59 reed      Reed Wade, Catalyst, NZ
09:59 irmaB     irma birchall from CALYX in Sydney
09:59 * clrh    Claire Hernandez, Biblibre
09:59 thd       Thomas Dukleth, Agogme, New York City
09:59 * hdl     Henri-Damien LAURENT, BibLibre
09:58 hdl       let's proceed to a round call
09:58 hdl       Hi..
09:58 irmaB     Hi all
09:58 munin     Useful Commands: #action #agreed #help #info #idea #link #topic.
09:58 munin     Meeting started Wed Dec 15 10:00:17 2010 UTC.  The chair is hdl. Information about MeetBot at http://wiki.debian.org/MeetBot.
09:58 hdl       #startmeeting solr meeting
09:57 reed      hellow
09:57 hdl       hi reed
09:56 hdl       meeting in two minutes.
09:55 magnus    hi Joubu
09:55 Joubu     hello
09:55 clrh      he is working with me on solr dev for biblibre
09:55 clrh      hey Joubu
09:48 hdl       yes tajoli
09:46 magnus    in about 10 minutes
09:46 clrh      hi
09:46 magnus    tajoli: yep
09:46 clrh      yep tajoli
09:46 tajoli    The Solr meeting will be here ?
09:45 tajoli    Hi to all
09:41 BobB      Guten arbend/morgen KF
09:41 chris     yep, just messaged you
09:41 kf        hi Irma and Bob
09:41 BobB      Hi Chris.  Can we have a chat somewhere?
09:40 chris     i know enough about dspace to know i dont like it
09:40 magnus    are you thinking of connecting them?
09:39 BobB      I need to find out who in the Koha community has lots of experience with DSpace?
09:38 BobB      That too.:)
09:38 magnus    yeah, not to mention irc nicks! ;-)
09:37 BobB      It was so good to meet people at KohaCon.  KohaCon turns email addresses into people!
09:36 magnus    BobB: not quite summer, but yes the return trip was as smooth as the outward trip
09:35 magnus    hehe
09:35 BobB      Summer!
09:35 munin     magnus: The current temperature in Bodo, Norway is 6.0�C (10:20 AM CET on December 15, 2010). Conditions: Mostly Cloudy. Humidity: 66%. Dew Point: 0.0�C. Windchill: 0.0�C. Pressure: 29.77 in 1008 hPa (Falling).
09:35 magnus    @wunder bodo, norway
09:34 BobB      Hi Magnus.  Obviously you arrived home safely from Wellington.  I hear it is very cold in Northern Europe.  It was 28 c in Sydney today.
09:34 chris     hi BobB
09:33 hdl       She is at the moment mostly on the solr developments.
09:33 magnus    hi clrh, nice to meet you! ;-)
09:33 clrh      thanks
09:33 BobB      Ahhh,  bienvenue!  Good to meet you.
09:33 hdl       clrh: is from BibLibre, our new dev manager.
09:32 clrh      Hi BobB :)
09:32 magnus    hi BobB & irmaB
09:32 BobB      Hi clrh.  I don't think we've met.  I'm Bob Birchall from Sydney Australia.
09:32 hdl       what about you BobB?
09:32 hdl       busy, as we all are.
09:31 clrh      It would be great for us to have more than 2 connected persons to katipo for one ip... si ?
09:31 BobB      I'll wait.  I have something to ask him.
09:30 hdl       (or was a few minutes ago)
09:30 BobB      Bon soir HDL, ca va?
09:29 hdl       hi BobB yes he is
09:28 BobB      Good evening all.  Is Chris Cormack about?
09:26 magnus    ok, will do
09:23 chris     yeah that would be good
09:23 magnus    should i take the logrotate stuff out and submit a new patch without it?
09:22 magnus    sounds nice, and a bit above me... ;-)
09:22 chris     that way, it can get installed with the package
09:22 chris     is probably a better fix and then making a file that they can drop into there
09:21 chris     with /etc/logrotate.d/
09:21 chris     and telling people to do it properly
09:20 chris     actually i think taking that line out altogether
09:19 magnus    probably ;-)
09:19 chris     ahh the logrotate.conf
09:18 chris     hmm i wonder what files i was going to create?
09:17 chris     lemme look what line 79 is
09:16 chris     not ringing a bell
09:07 magnus    thanks, clrh
09:07 munin     04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=5055 enhancement, PATCH-Sent, ---, magnus, NEW, crontab.example should use standard file paths
09:07 magnus    chris: did you ever get around to looking at the logrotate stuff i mention you mentioning in bug 5055?
09:07 clrh      right magnus
09:06 magnus    hey chris
09:06 chris     hi magnus
09:06 magnus    solr meeting in about 50 minutes, right?
09:06 magnus    kia ora, #koha!
09:00 chris     hi clrh
08:56 clrh      Hello everyone
08:40 kf        hi ivanc
08:35 kf        yeah, no problem. :) Just eager to start.
08:35 fredericd kf: yes, I'm updating all .po files and try to automate the process
08:35 ivanc     hi kf
08:34 ivanc     hi #koha
08:34 kf        fredericd: can you ping me when pootle is back?
08:32 kf        but lot's of new strings I expect
08:31 fredericd yes, it could have been worse
08:30 kf        could be worse :)
08:30 kf        puh
08:30 fredericd kf: for German you will 300 string exactly marked as fuzzy
08:30 kf        ok
08:30 kf        ah
08:29 kf        you don't really have to reatranslate, but submit them again
08:29 fredericd it was my fault
08:29 kf        I think strings went fuzzy before when the template file got touched
08:29 fredericd I was discovering that the update was marking as fuzzy more than 1800 string in French
08:29 kf        ah
08:29 fredericd kf: it was about .po files update for 3.2.2
08:24 kf        sorry, missed the beginning of your conversation
08:24 kf        fredericd: are you updating the pootle files today?
08:24 hdl       fredericd: pleased for you
08:23 kf        hi fredericd
08:22 fredericd hi kf
08:22 fredericd I was working on hdl .po files version, before updates we've done the last few days
08:22 kf        hi #koha
08:21 fredericd I wasn't working on the correct French version of .po files
08:21 fredericd I was wrong for the fuzzy strings in French
08:21 fredericd YEAH
08:15 chris     once for each release is enough
08:15 fredericd I'm not sure it will be possible to update translation from 3.2.x regularly as planned
08:14 chris     theres certainly no way we are going to try to bend it to work with TT
08:14 fredericd hope so
08:14 chris     just for 3.2.x
08:14 chris     e
08:14 chris     yeah, it doesnt have too much longer to liv
08:14 fredericd tmpl_process3.pl is a blackbox for me, I can't fix it
08:14 chris     i reckon go from about 2000 to about 200 lines
08:13 chris     to extract the alt tags etc
08:13 chris     so you can just use a proper HTML parser
08:13 chris     easy
08:13 chris     [% %]
08:13 chris     that arent really comments
08:13 chris     we dont have to try and look for comments
08:12 chris     its already underway
08:12 chris     yep, its tons easier
08:12 fredericd Clearly the templates strings extraction will have to be improved for TT
08:11 fredericd 300 strings in German, so there is something special in French
08:11 hdl       it does require work
08:09 hdl       welcome the translating process ;)
08:09 fredericd hdl: Yes, but 1843 strings...
08:08 hdl       fredericd: but some will still need some work
08:07 hdl       fredericd: most of those strings will be unchanged. and will just have to be confirmed...
08:05 chris     and that will be msmerge doing that
08:04 chris     yep, that would make them fuzzy
08:04 hdl       fredericd: it can be that strings changed in context
08:04 fredericd but less fuzzy: just 356 strings
08:03 chris     and it certainly never used to do this
08:03 chris     hmm, afaik nothing has changed with tmpl_process3.pl
08:03 fredericd chris: same for en_NZ
08:03 fredericd I don't think so, because the French .po file have been update outside Pootle
08:02 chris     i wonder if its to do with new pootle?
08:02 chris     without that happening
08:02 chris     i had updated them quite a few times
08:01 chris     because that had nothing untranslated
08:01 fredericd yes, I desactivated Pootle in the meantime
08:01 chris     you could try updating english-nz
08:01 fredericd just before, the French .po files were without fuzzy
08:01 chris     is pootle down?
08:01 Oak       \o
08:01 fredericd chris: yes
08:01 fredericd new string are added, which is the expected behavior
08:00 chris     and they definitely werent fuzzy before?
08:00 fredericd yes, just an update
08:00 chris     when you did an update?
08:00 fredericd for example on French staff .po files 1843 strings have been marked as fuzzy
07:59 chris     hmmm interesting that must be a new thing, because it didnt used to
07:58 fredericd on stable branch, it means that to be able to have a few new strings translated into your language, you have to do a huge work of re-translation...
07:56 fredericd it means that translators will have to re-translate string they already translated
07:56 fredericd A LOT of string are marked as fuzzy, but they shouldn't
07:55 fredericd chris: Are you aware that tmpl_process3.pl script mess up templates .po files?
06:30 cait      hmpf!
06:28 * Brooke  cites the old adage that good things happen early in the morning.
06:27 Brooke    eeyah
06:27 cait      morning Brooke
06:27 * Brooke  waves at Chris
06:27 cait      morning chris
06:27 chris     hi cait and Brooke
06:22 cait      hi #koha
04:56 kmkale    Good morning
04:54 * munin   reloads and spins the chambers.
04:54 Brooke    @roulette
03:41 chris     night
03:40 chris_n   off to sleep, g'night
03:37 chris_n   scarry thought :)
03:36 chris     http://xkcd.com/224/
03:36 chris     of course
03:35 chris_n   in perl... of course :-)
03:35 chris     cool :)
03:35 chris_n   its pretty cool; it texts the temps in the loop along with the boiler status
03:35 chris     will do :)
03:35 chris_n   tell sharrow that I put together a script to have nagios monitor the hot water boiler loop in a large chilled water cooling system this week
03:33 chris     ahhh right
03:33 chris_n   end of fall semester schedule pains
03:33 chris     oh yeah?
03:33 chris_n   its been one crazy ride the past two weeks here :-P
03:32 chris     :)
03:32 chris_n   pushed even
03:32 chris_n   pused
03:32 chris     cool
03:32 chris_n   I did
03:32 * chris_n just wandered in
03:31 chris     yup
03:31 chris_n   chris: was that directed to me?
03:24 chris     http://git.koha-community.org/gitweb/?p=koha.git;a=commit;h=3705965cd71bc30ed4ba325cde115764fa1afe44
03:23 chris     did you catch the revert for Ergonomy improvement in smart rule management
03:23 chris     chris_n`: you about?
03:22 hudsonbot * Katrin Fischer: Bug 2965: Allow due date in the past - small template fix
03:22 hudsonbot * Robin Sheat: Bug 5084 - hide funds that are part of an inactive budget
03:22 hudsonbot * Srdjan Jankovic: Bug 2965: Allow due date in the past
03:22 hudsonbot * Owen Leonard: Fix for Bug 5000, Uncertain prices misses option to choose display language
03:22 hudsonbot * Colin Campbell: Variable redeclared in same scope
03:22 hudsonbot * Nicole Engard: bug 5255 change 'document type' to 'item type'
03:22 hudsonbot * Nicole Engard: bug 4252 add authorites permission to menus
03:22 hudsonbot * Colin Campbell: Bug 2170 Supplementary Fix Wrap link in permissions check
03:22 hudsonbot * Owen Leonard: Updated fix for Bug 2170, Adding 'edititems' user-permission
03:22 hudsonbot * Chris Cormack: Bug 5484 - Handling bad borrower categories in serial routing lists more gracefully
03:22 hudsonbot * Nicole Engard: bug 5150 change issuing to circ & fine rules
03:22 hudsonbot * Owen Leonard: Fix for Bug 5208, Language chooser missing on Batch item deletion/modification
03:22 hudsonbot * Robin Sheat: Bug 5228 - make rebuild_zebra handle fixing the zebra dirs
03:22 hudsonbot * Robin Sheat: Bug 5313 - allow creation of libraries with hyphens
03:22 hudsonbot * Owen Leonard: Fix for Bug 5416, Template syntax error in moredetails.tmpl
03:22 hudsonbot * Frédéric Demians: Bug 5041 Allow to delete non-repeatable field
03:22 hudsonbot * Katrin Fischer: Bug 4218: Fixes display problem introduced by last patch
03:22 hudsonbot Project Koha_3.2.x build #59: STILL UNSTABLE in 22 min: http://hudson.koha-community.org/job/Koha_3.2.x/59/
03:00 hudsonbot Starting build 59 for job Koha_3.2.x (previous build: STILL UNSTABLE -- last SUCCESS #54 8 days 4 hr ago)
02:56 druthb    hi, chris!  :)
02:56 chris     hi druthb
02:50 munin     New commit(s) kohagit32: Fix for 5143, now with IE debug removed <http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=92070b30398e8700a2db6db8b44c9a8f740311b7> / Bug 3789 Set off shelving location in staff and OPAC title display <http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=d19300432fe0f7a940e7b685a343b7e505bbd0d4> / Bug 4937: Fixes XHTML in the pagination links of a saved report. <http://git.koha-community.org/gitwe
00:57 ebegin    Is there a way to know what is indexed in zebra for a specific bib?
00:55 sylvar    OK, thanks.
00:55 chris     yes
00:54 sylvar    OK, I thought I knew this, but... is borrowers.cardnumber where barcodes should go?
00:37 ebegin    I'm having trouble on my side to find word preceded of "L'", for exemple, I can find "L'entreprise" but not just "Entreprise"...
00:24 ebegin    or "kw,wrdl: your search"
00:23 ebegin    in your breadcrumbs, what koha shows? "kw,phr: your search"
00:23 sylvar    Thanks, though.
00:23 chris     :)
00:22 sylvar    chris: 3.02.00.007, but I suspect I've made a mistake during the data load, so I wouldn't consider it a bug.  Not a Koha bug, anyhow. ;)
00:19 chris     hm interesting, what version of koha?
00:19 sylvar    I'll double-check.  The bib detail page says they're all Available, on both sides.
00:19 chris     none of them are marked lost?
00:18 sylvar    Hi y'all. If a search from the admin side shows 14 items, 14 available (for a particular bib record) and the same search phrase on the public side shows "(1)", what might cause that difference?