Time  Nick              Message
20:14 jcamins           That's basically just a general rule.
20:13 talljoy_tickets   HA
20:13 jcamins           Ah, yeah, you're SOL when it comes to 773s.
20:12 talljoy_tickets   and shoving 773 tags into those.
20:12 talljoy_tickets   mainly my issue is with host items (analytics).  and i think this is going to be a custom export using the C1, C2 and C3 RIS fields.
20:11 talljoy_tickets   i've found the koha code.  but before i add more, i thought it would be a good idea to see if anyone cared what i shoved into Alternate Name.  heh
20:11 jcamins           I believe the code is somewhere in C4.
20:11 talljoy_tickets   yup i'm digging around and it seems 'loosey goosey'
20:10 jcamins           talljoy_tickets: you might take a look at marc2bibtex.
19:37 talljoy_tickets   i know it's friday, but is anyone still in that knows if there are any 'rules' for converting marc21 to Bibtex citations?  i.e. what tag goes into what output field
19:21 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=12065 minor, P5 - low, ---, m.de.rooy, Pushed to Master , Consistent use of encode_qp when sending basket/shelf
19:21 jenkins_koha      Marcel de Rooy: Bug 12065: use encode_qp consistently when sending basket/shelf
19:21 jenkins_koha      Project Koha_master build #1751: SUCCESS in 2 hr 2 min: http://jenkins.koha-community.org/job/Koha_master/1751/
18:57 tcohen            bye #koha
18:46 tcohen            i've just noticed we don't have it, but have '01e' instead
18:46 tcohen            is field 019 of any use in MARC21?
18:45 markvandenborre   bye
18:45 markvandenborre   thx all for the help
18:45 markvandenborre   have to run
18:45 markvandenborre   I mapped correctly now for unimarc, but still no cigar
18:41 markvandenborre   let's hope that is the case...
18:41 markvandenborre   ah, it only becomes searchable when the koha mapping biblio.title and biblio.author are filled in maybe?
18:33 markvandenborre   koha is so much more heavyweight than what we actually need, but I _will_ kick it into submission
18:32 markvandenborre   I'd hoped to have something running by now
18:31 markvandenborre   sorry about all these questions, all...
18:26 markvandenborre   (so that I can test while certain)
18:26 markvandenborre   what is the cleanest way to remove all existing records from koha?
18:24 markvandenborre   the intelligent search in the koha interface does help though
18:19 markvandenborre   like title, author, ...
18:19 markvandenborre   if only there were a way to find out the number for the ten most important bits of information
18:18 markvandenborre   aargh, this is such a long list of meta-metadata
18:15 jcamins           Though I guess you weren't getting permissions issues.
18:15 markvandenborre   yup
18:15 jcamins           Then you don't have to worry about it connecting to the database.
18:15 jcamins           Actually, just use the MARC field.
18:14 jcamins           Title should be fine.
18:14 markvandenborre   together with the title one maybe
18:14 markvandenborre   so that may be the issue
18:14 markvandenborre   that one was the only one that I took without double checking
18:13 markvandenborre   ah, that is the one before indeed
18:13 jcamins           100$a is a required UNIMARC field, and it doesn't have composer in it.
18:13 markvandenborre   looking at my koha test instance for reference
18:13 markvandenborre   and I used those
18:12 markvandenborre   ok, but I checked the equivalent unimarc fields
18:12 jcamins           Everyone here uses MARC21, so those are the fields we suggested.
18:11 markvandenborre   I'm sorry, I'm a bit confused now
18:05 wahanui           unimarc is http://www.ifla.org/en/publications/unimarc-formats-and-related-documentation
18:05 jcamins           unimarc?
18:04 jcamins           No, all the MARC fields.
18:04 markvandenborre   you mean for the title field?
18:03 jcamins           You'll need to identify appropriate UNIMARC fields to replace the fields.
18:03 jcamins           Yeah, those are MARC21 fields.
18:02 markvandenborre   was my exact line
18:02 markvandenborre   ./csvtomarc.pl --input=/root/metadata_itunes_geraardsbergen.csv --output=/root/bib_gbergen.marc --mapping='titel=title' --mapping='componist=marc:100_a?' --mapping='code=marc:500_2?' --mapping='instrument=marc:500_r?' --mapping='graad=marc:500_n?' --mapping='type=marc:500_x?' --kohaconf=/etc/koha/sites/gbergen/koha-conf.xml --kohalibs=/usr/share/koha/lib --fieldsep=';'  --format=usmarc
18:02 jcamins           markvandenborre: what field did you put the title in?
18:02 markvandenborre   but nothing yet in the web frontend
18:02 markvandenborre   on the rebuild zebra command line
18:01 markvandenborre   it seems to find all records
18:01 markvandenborre   http://pastie.org/9160037
17:58 gmcharlt          try a "sudo koha-rebuild-zebra -b -f -v instancename"
17:57 gmcharlt          right
17:57 markvandenborre   ok
17:57 markvandenborre   ah, it doesn't rely on zebra?
17:57 gmcharlt          yeah, that's a separate search mode
17:57 markvandenborre   but using the normal "search" not
17:57 markvandenborre   that yields some results
17:57 markvandenborre   "cataloging search"
17:57 markvandenborre   as I said, I can search the catalog using
17:56 markvandenborre   yes
17:56 gmcharlt          then let's see if the search engine is running -- is there at least one process called zebrasrv running?
17:55 markvandenborre   it seems to be indexing in some way
17:55 markvandenborre   gmcharlt: I installed from debian packages
17:55 gmcharlt          if not, you'd need to run rebuild_zebra
17:54 gmcharlt          markvandenborre: you need to index; if you installed from packages, the relevant cronjob should already be set up
17:51 * markvandenborre reading the manual again
17:47 markvandenborre   but search not
17:47 markvandenborre   "cataloging search" works
17:47 markvandenborre   I was too quick thinking that
17:44 markvandenborre   ah, a restart of the catalog helped
17:44 markvandenborre   I feel like I am missing something vitally important here, and I have no idea what
17:41 markvandenborre   do I need to trigger a kind of indexing in any way?
17:41 markvandenborre   is there anything obvious I am missing?
17:40 markvandenborre   status for every record says "imported"
17:38 markvandenborre   which fails... obviously, I'm missing something here
17:38 markvandenborre   so the first thing I try to do is search for them of course
17:37 markvandenborre   and added them to the catalog (I think)
17:37 markvandenborre   ok, so now I have staged 2702 records for import
17:33 tcohen            heh
17:32 jcamins           tcohen: yes, but I will never ever suggest someone use bulkmarcimport because I don't want to be responsible for making someone use bulkmarcimport. :P
17:32 tcohen            bulkmarcimport accepts MARCXML, doesn't?
17:31 jcamins           Right, unless you want to load it into Koha.
17:31 cait              marcxml is way easier to read for testing
17:31 jcamins           However, if you're using UNIMARC, you'll want to consult the documentation for UNIMARC to figure out what fields to use. I don't know UNIMARC.
17:30 jcamins           (which is the name of the binary format that is used for both MARC21 and UNIMARC)
17:30 markvandenborre   ok
17:30 jcamins           Koha requires usmarc.
17:30 markvandenborre   and B<marcxml>. B<marcxml> is the default.
17:30 markvandenborre   This is the file format that will be output. Valid options are B<usmarc>
17:30 markvandenborre   this is what the csvtomarc.pl manpage says:
17:30 markvandenborre   I wonder if that has something to do with the marc dialect
17:29 markvandenborre   it only imported one record
17:29 markvandenborre   I tried to import what I had reformatted into a running koha instance
17:28 huginn            khall: The operation succeeded.
17:28 khall             @later tell cait I've left a comment/question for you on bug 9303
17:20 jenkins_koha      Starting build #1751 for job Koha_master (previous build: SUCCESS)
17:16 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=12214 minor, P5 - low, ---, pasi.kallinen, Pushed to Master , SQL errors in reports are not shown to user
17:16 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=11975 enhancement, P5 - low, ---, jonathan.druart, Pushed to Master , Tools: Batch patron deletion code should be improved
17:16 jenkins_koha      * Kyle M Hall: Bug 12214: (follow-up) Clean up reports-toolbar.inc, show Edit link when SQL has errors
17:16 jenkins_koha      * Galen Charlton: Bug 12214: (follow-up) correct POD of C4::Reports::Guided::execute_query()
17:16 jenkins_koha      * Pasi Kallinen: Bug 12214: display SQL errors in reports to users
17:16 jenkins_koha      * Galen Charlton: Bug 12214: add regression test for reporting error when running report with SQL error
17:16 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=11843 blocker, P5 - low, ---, julian.maurice, Pushed to Master , Manual subscription history doesn't seem to work as expected
17:16 jenkins_koha      * Galen Charlton: Bug 11975: (follow-up) simplify construction of params for GetBorrowersToExpunge()
17:16 jenkins_koha      * Jonathan Druart: Bug 11975: improve the batch patron deletion code
17:16 jenkins_koha      * Julian Maurice: Bug 11843: (follow-up) fix unit test in t/db_dependent/Serials.t
17:16 jenkins_koha      * Julian Maurice: Bug 11843: prevent manual history from being overwritten if subscription switched to automatic history
17:16 jenkins_koha      Project Koha_master build #1750: SUCCESS in 2 hr 17 min: http://jenkins.koha-community.org/job/Koha_master/1750/
17:08 chris_n           strawberries, even
17:08 chris_n           had strawberrys, biscuit, and whipped cream for breakfast
17:07 markvandenborre   some really good wine too!
17:07 markvandenborre   and (very few people seem to know this)
17:07 markvandenborre   and beer
17:07 markvandenborre   and cheese
17:07 markvandenborre   and strawberries, this time of the year
17:06 jcamins           chris_n: ooh, I have paneer saag for lunch. I think I'll go eat.
17:06 markvandenborre   chris_n: and chocolate, and witloof
17:05 * chris_n         begins to think in terms of pastry and icecream
17:05 jcamins           Ah, yeah, that might be. You're in Belgium?
17:05 markvandenborre   but if most people here use the other standard, that's worth a lot to me
17:05 markvandenborre   I read that Belgium was more into the French flavor...
17:04 gmcharlt          but the two main flavors are UNIMARC and MARC21
17:04 jcamins           France/Portugal/Spain vs. the rest of the world.
17:04 gmcharlt          it's not a strict US vs EU distinction
17:04 gmcharlt          yes
17:04 markvandenborre   different Marc dialect?
17:04 markvandenborre   ?
17:04 markvandenborre   versus US
17:04 markvandenborre   ah, EU
17:03 huginn            gmcharlt: General note
17:03 gmcharlt          @marc 500 a
17:03 huginn            gmcharlt: unknown tag 500$a
17:03 gmcharlt          @marc 500$a
17:03 jcamins           markvandenborre: where are you located?
17:03 gmcharlt          500a
17:03 markvandenborre   gmcharlt: you mean 500n (misc info) instead of 500a (uniform title) I presume?
17:01 markvandenborre   thx
17:01 markvandenborre   and saw what you meant
17:01 markvandenborre   I just dug in more deeply
17:01 jcamins           markvandenborre: I meant there was a good chance a field would already exist, and if not, you can use a related field, or, if there isn't even one that's similar, 500 is good for everything.
17:00 ashimema          :)
17:00 markvandenborre   cool
17:00 ashimema          perls still alive and kicking.. but feel free to re-wrtie koha in python if you have the inclination ;)
17:00 gmcharlt          the 500a (general note) is a good catchall
16:59 markvandenborre   (or abuse an unrelated field)
16:59 markvandenborre   jcamins: or do you mean like that is a place where one could use an unrelated field
16:58 talljoy           happy to help markvandenborre
16:57 * markvandenborre has looked at 5xx and not really found anything useful there
16:56 oleonard          Then computers are ancient too, but that doesn't mean you have to use the oldest one.
16:56 markvandenborre   also, about really nice people like jcamins or cait or talljoy or ...
16:56 * markvandenborre ducks
16:56 markvandenborre   ...and perl
16:55 oleonard          ...assuming you're talking about MARC and the holds system.
16:55 jcamins           That's pretty accurate.
16:55 markvandenborre   koha really feels like a very friendly community around really ancient technology
16:54 jcamins           Okay.
16:53 markvandenborre   jcamins: I had already figured that out more or less
16:53 jcamins           markvandenborre: there may be a field for that sort of thing. There probably is, in the 5xx fields.
16:52 bag_away          oleonard++
16:52 oleonard          We should really stop letting people believe that Koha can restrict the number of holds a patron can place.
16:52 markvandenborre   so let me take an example... instrument
16:52 jcamins           The MARC standard should have been retired >20 years ago. Just making it work is enough trouble.
16:52 * jcamins         takes a very cavalier attitude toward these things.
16:51 jcamins           If you're not a library cataloger, you won't care, so why would anyone else?
16:51 markvandenborre   ?
16:51 markvandenborre   ...but that would rather be abusing fields
16:50 jcamins           You can use any fields for local information, if you're not sharing your records.
16:50 markvandenborre   unless I create items then, right?
16:50 markvandenborre   ...and presumably, there is no way to keep that local information
16:49 jcamins           Well... usually.
16:48 jcamins           Exactly.
16:48 markvandenborre   ah, an item is a local "instance" of a particular book?
16:47 jcamins           Well, that could be an argument for creating items, so you could use the call number field.
16:47 markvandenborre   it wasn't applied to every piece of sheet music
16:47 markvandenborre   applied locally to the book
16:47 markvandenborre   the code is just a label that was (inconsistently)
16:47 jcamins           markvandenborre: there's pretty extensive Perldoc, as I recall. But the actual fields you'll have to figure out by talking to your catalogers (or if you're the cataloger, by looking at the standard).
16:46 markvandenborre   how would you map an existing unique code to marc?
16:40 markvandenborre   by RTFS (reading the ....... source)
16:40 markvandenborre   talljoy: thx, just found that out too
16:25 talljoy           SHOW is the column heading and it is mapped to the marc tag 245 a.
16:24 talljoy           --mapping=SHOW=marc:245_a?
16:24 talljoy           this is an example of one of my mapping lines in the csvtomarc script
16:24 markvandenborre   :-)
16:24 markvandenborre   or, to be correct, German sheet music records
16:24 cait              nice
16:24 markvandenborre   German sheet music
16:24 cait              german records? :)
16:23 markvandenborre   marc:100a
16:23 markvandenborre   correct, just found information about that in the source of the script
16:23 cait              i think you need to tell it the field it goes to - 245a somewhere
16:23 gmcharlt          could you point to the one you're using specifically?
16:23 gmcharlt          hmm; there are a few scripts of that name running around
16:22 markvandenborre   and running csvtomarc.pl with only the 'titel=title' mapping results in something like http://pastie.org/9159718
16:21 markvandenborre   somehow, this script "knows" what "title" means
16:19 oleonard          khall: I think this is incorrect.
16:19 markvandenborre   for use with this koha migration toolbox script csvtomarc.pl
16:19 oleonard          khall: So online registration requires that  autoMemberNum be turned on, otherwise patrons who register online will be created with no barcode number
16:19 markvandenborre   so I'm not entirely sure yet where I can find a list of destination mappings
16:18 markvandenborre   is what I get...
16:18 markvandenborre   100a doesn't appear to be a valid Koha field name.
16:17 markvandenborre   ok
16:16 markvandenborre   as mappings in csvtomarc.pl?
16:16 markvandenborre   ?
16:16 markvandenborre   so I just use these numbers that I can find in the url that talljoy just gave
16:14 talljoy           additional composers/lyricists 700a
16:14 gmcharlt          composer = 100a
16:14 gmcharlt          title = 245a
16:14 gmcharlt          then for the easy ones
16:13 markvandenborre   gmcharlt: all sheet music, correct
16:13 talljoy           http://oclc.org/bibformats/en/0xx.html   is a great resource for marc tags
16:13 talljoy           yes your url woud go in th 856u
16:13 gmcharlt          are your records all works of musics?
16:12 markvandenborre   that stands for url
16:12 markvandenborre   and found something like 856u there
16:12 talljoy           oh, that is cataloging question.
16:12 markvandenborre   I have a url field and I have been looking into koha
16:12 markvandenborre   but I don't know the name of the marc field to map this to
16:11 markvandenborre   header in the csv file
16:11 markvandenborre   so I have a composer
16:11 markvandenborre   but I don't know the exact name of the marc side of mappings of the fields
16:11 markvandenborre   I understand the syntax, and it's working correctly
16:11 talljoy           markvandenborre i've used it before and ethyian is the one who wrote it, if i'm not mistaken.
16:10 markvandenborre   but I'm not 100% where to find this
16:10 markvandenborre   the csvtomarc.pl tool gives the example of title as something to map
16:10 markvandenborre   title, code, level (string), various (string), instrument, composer and (internal) url
16:09 markvandenborre   how to map these very basic fields from csv to marc
16:09 markvandenborre   I'm looking through all kinds of metadata fields and I'm not really sure
15:24 huginn            New commit(s) kohagit: Bug 12065: use encode_qp consistently when sending basket/shelf <http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=76e8bb0fabd4a9ef85fd2dc5ad604c446a50f751>
15:23 markvandenborre   it seems to be workign
15:23 markvandenborre   thx for the hint
15:23 markvandenborre   just did that
15:21 jcamins           Yeah, that's why the best thing to do is change the separator character the script is using.
15:20 markvandenborre   I tried to avoid using commas since some book titles tend to contain them...
15:18 jcamins           CSV is generally pronounced "comma-separated values," even though technically it stands for "character-separated values."
15:17 markvandenborre   brilliant, thank you
15:17 markvandenborre   _ah_
15:17 jcamins           Comma.
15:17 markvandenborre   jcamins: I was hoping someone would be able to tell me what separator it was expecting
15:17 markvandenborre   it doesn't seem to tell what separator it expects, but indeed...
15:16 jcamins           markvandenborre: I can't really troubleshoot, but if I had to guess, I'd say you might not have changed the separator character.
15:16 markvandenborre   I'm not a perfect perl reader
15:16 markvandenborre   jcamins: any idea?
15:15 markvandenborre   code;titel;graad;stuk/studie/methode (string);instrument;componist (string);
15:15 markvandenborre   while the first line of that file says:
15:15 markvandenborre   Missing column specified in the mapping: titel doesn't exist in migrate.csv
15:14 markvandenborre   but for some reason it complains about
15:14 markvandenborre   I have all dependencies installed, and I am quite certain I am calling it in the right way
15:14 markvandenborre   hm, anyone used csvtomarc.pl before?
15:06 cait              hi tcohen
15:05 reiveune          bye
15:05 tcohen            hi cait
15:02 cait              nice
15:01 jenkins_koha      Starting build #1750 for job Koha_master (previous build: SUCCESS)
14:54 huginn            New commit(s) kohagit: Bug 12214: (follow-up) Clean up reports-toolbar.inc, show Edit link when SQL has... <http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=31781ffb02c3a806e0956ca07dfc24beab0b88dd> / Bug 12214: (follow-up) correct POD of C4::Reports::Guided::execute_query() <http://git.koha-community.org/gitweb/?p=koha.git;a=commitdiff;h=d514c21aed640be48dfbc70d3976ec0f865bb370> / Bug 12214: display SQL errors in reports to users
14:45 huginn            rambutan: The current temperature in Wyatt Park, St Joseph, Missouri is 13.4°C (9:45 AM CDT on May 09, 2014). Conditions: Clear. Humidity: 64%. Dew Point: 7.0°C. Pressure: 29.91 in 1013 hPa (Rising).
14:45 rambutan          @wunder 64507
14:19 * mveron          ...thinks that it could be set to critical or higher...
14:17 oleonard          khall: And then you can sign off :)
14:16 mveron            Sorry, mistaked, has "Needs Signoff" as status.
14:14 khall             thanks!
14:13 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=9612 minor, P5 - low, ---, m.de.rooy, Needs Signoff , SRU Response is different when DOM indexing is enabled
14:13 mveron            khall: See Bug 9612 - SRU Response is different when DOM indexing is enabled . A patch  that fixes this issue is pushed.
14:05 khall             I'm getting bad search results on master: http://i.imgur.com/ff3BorL.png
13:38 markvandenborre   jcamins: thank you for the hints
13:38 huginn            mveron: The current temperature in Wetter Allschwil, Allschwil, Switzerland is 22.1°C (3:38 PM CEST on May 09, 2014). Conditions: Mostly Cloudy. Humidity: 55%. Dew Point: 13.0°C. Pressure: 30.06 in 1018 hPa (Steady).
13:38 mveron            @wunder allschwil
13:38 mveron            Good afternoon barton :-)
13:36 barton            morning mveron
13:36 barton            ashimema: I just made those changes; also I added a 'REGENERATE THE DOM CONFIGS' step -- there should be a link there to any documentation that we have about that.
13:35 mveron            Hi #koha
13:32 jcamins           You'll have to read the documentation for that script, because you have to do all the mapping on the command line.
13:31 jcamins           It generates MARC records which you can load through the Stage MARC Record tool.
13:31 markvandenborre   jcamins: so I run the csvtomarc.pl script, and I get a text file that I can import somehow into koha...
13:21 markvandenborre   jcamins: /me is looking at the conversion scripts
13:20 markvandenborre   and we don't want the physical copies of the sheet music to be used
13:20 markvandenborre   the main one
13:20 markvandenborre   there is only one branch
13:20 markvandenborre   jcamins: no, we don't want that (for the moment)
13:19 jcamins           But whether that's what you want to do is a policy decision. If you have multiple branches, or the sheet music exists somewhere physically as well, you might want items.
13:19 markvandenborre   the point of course being that all of these will need to be edited (pdf url field needs to be customised)
13:19 markvandenborre   ok, I'll try and do that first then
13:18 jcamins           I would not add any items.
13:18 jcamins           If I had scanned sheet music, all I would do is convert the CSV into MARC and load it into Koha.
13:18 markvandenborre   there seem to be quite a few layers of metadata in koha (probably rightly so for many situations)
13:17 markvandenborre   jcamins: I was hoping to get a better idea of minimal steps to take
13:16 jcamins           Possibly. That's a policy decision. How you want to handle item types, etc.
13:14 markvandenborre   but don't I need to define an item type for this kind of thing first?
13:13 wahanui           oleonard: I forgot in that case
13:13 oleonard          wahanui: forget in that case
13:13 jcamins           ^^ there's a csv to marc script in the migration toolbox.
13:12 wahanui           well, migration toolbox is https://gitorious.org/koha-migration-toolbox
13:12 jcamins           migration toolbox?
13:12 jcamins           toolbox?
13:12 jcamins           koha toolbox?
13:12 jcamins           koha tool box?
13:12 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=11631 enhancement, P5 - low, ---, pasi.kallinen, Pushed to Master , Make translation toolchain ignore useless translatable strings
13:12 wahanui           in that case is preferable thar each text be inside some tag, then it will be picked alone, and not with all %s mess, they are left and  removed by Bug 11631
13:12 jcamins           markvandenborre: in that case...
13:12 markvandenborre   jcamins: I have pulled some csv out of iTunes (don't ask how, it was horrible)
13:11 jcamins           There is already a script for converting CSVs to MARC.
13:10 jcamins           markvandenborre: you'll need to convert your records into MARC format. You'll probably need to write a script to do that, though if you can get a CSV out of iTunes, you might be able to make use of that.
13:10 markvandenborre   asap
13:10 markvandenborre   but to keep things as simple as possible in the short term, and to have something up and running
13:09 markvandenborre   the goal would be to upgrade this data set in the long term with some extra fields (publisher, ...)
13:08 markvandenborre   without wading through piles and piles of metadata fields?
13:08 markvandenborre   is there an easy and clean way to import these into koha?
13:07 markvandenborre   and pdf document
13:07 markvandenborre   the only fields are: title, code, level (string), various (string), instrument, composer
13:07 markvandenborre   right now, the only metadata we have is in iTunes (don't ask)
13:06 markvandenborre   I have a library of 2702 pieces of sheet music
13:03 barton            Yeah, I will, once I've been through it once or twice. I'm still learning...
13:02 ashimema          I left it under my username whilst it was a bit crude.. but it sounds like you've got allot of it tidied up nicely now..
13:01 ashimema          feel free to 'publish' it somewhere more sensible if you like.
13:01 ashimema          :)
13:01 ashimema          it's a while since i've actually followed the instructions to be honest.. I understand it just about well enough to fly by the seed of my pants now ;)
13:01 barton            i'll update your wiki page.
13:00 ashimema          sounds fair to me barton
13:00 barton            I think that it's worth doing a diff after the fact, just to make sure...
12:59 barton            ashimema: If you fix the ~/kohaclone/etc/zebradb/marc_defs/marc21/biblios/biblio-koha-indexdefs.xml before you run xsltproc, that basically takes care of ~/kohaclone/etc/zebradb/marc_defs/marc21/biblios/biblio-zebra-indexdefs.xsl ... the only changes that I had there were the ones that I wanted.
12:46 tcohen            hi magnuse
12:29 tcohen            hi ashimema
12:29 tcohen            dcookhi
12:28 ashimema          morning tcohen
12:28 tcohen            morning #koha
12:18 cait              hi oleonard :)
12:17 oleonard          Hi #koha
12:12 ashimema          glad to hear it ;)
12:12 ashimema          :)
12:10 jcamins           ashimema: yeah, there was a moment of Transendent Moment of Understanding.
12:10 ashimema          did I miss anything
12:10 ashimema          ack.. my IRC just crashed..
12:08 barton            ok, that makes sense, more or less.
12:05 ashimema          I'm sure it's not too bad.. but I've been trying to keep the configs in sync whilst we support both.. and i've found the easiest way was to regenerate and then diff accross the 'special' bits for dom
12:03 jcamins           gmcharlt did write some documentation, but I don't remember where it is. It's a pretty straightforward format.
12:03 barton            so ... is there someplace that I could read up on regenerating DOM configs?
12:00 ashimema          :)
12:00 jcamins           Agreed.
12:00 ashimema          which reminds me.. I'd like a aptch to ditch all the comments in those dom files at some point..  if w don't want them relating to grs anymore.. then they don't need to be there.. they're just confusing
12:00 barton            hm. that slightly rings a bell ...
11:59 ashimema          hehe.. yup.. that's how I remember it
11:58 jcamins           barton: really? I could've sworn you were part of the conversation when ashimena added the page and I said "STOP!!!! DON'T FOLLOW THOSE DIRECTIONS UNTIL YOU KNOW THAT YOU SHOULDN'T REGENERATE THE DOM CONFIGS!!!"
11:58 ashimema          I remember ;)
11:58 ashimema          he did.. when I first added that page to the wiki for you ;)
11:57 barton            um, jcamins ... actually I *don't* remember you polnting out to me that this was a bad idea...
11:57 ashimema          I really aught to take another look at that bug.. we simply can't use QP until that's fixed :'(
11:57 jcamins           You said it!
11:56 ashimema          hehe :)
11:56 ashimema          supression is plain baldy implimented
11:56 ashimema          eythian was having similar issues with getting elasticsearch working with supression
11:56 ashimema          I'm still stuck on getting QueryParser working with opacsupression though..
11:56 * barton          stands corrected.
11:55 jcamins           Also, I've already pointed out to Barton that doing that is a Bad Idea, so there's no one involved in the conversation who might leave with incorrect ideas.
11:55 ashimema          hehe.. thanks jcamins.. that's reasonable..
11:55 barton            maybe jcamins hasn't had his coffee yet, either.
11:55 jcamins           ashimema: if you want to do something stupid after you *know* it's stupid, I don't really care... it's just when there's a possibility that you're doing something dumb because you don't know any better that I'll point it out.
11:54 ashimema          jcamins, surprised you've not shot me down in flames yet ;)
11:54 barton            ashimema++
11:54 ashimema          but I'm no pro in writing the DOM config manually.. so I use them and deal with the consequences
11:53 ashimema          Jared still won't like them, as he goes on about never ever using those scripts to generate configs again..
11:53 * cait            prefer taming it
11:53 ashimema          ok, barton, I've updated the instructions a bit to be more generic
11:53 barton            (appoligies to elmer fudd)
11:52 barton            kill the zebra, kill the, kill the ZEEEE-BRA!!!
11:52 cait              good luck :)
11:52 barton            ok. I've got my spear and magic helmet ...
11:48 cait              ... it as just too tempting :)
11:48 ashimema          haha.. classic typo
11:48 ashimema          if you are, you'll deffo need a hacked up version of those additional steps ;)
11:48 barton            using yes, suing, no ;-)
11:48 cait              why would he sue the dom?
11:48 ashimema          Are you suing Dom barton?
11:47 cait              hi khall
11:47 khall             hi cait!
11:47 cait              barton: we also go the full package approach, i think there were some security concerns about git installs in the past
11:47 barton            hm.
11:46 ashimema          problem is.. those scripts a) create a massive diff due to the comments with linenumber, and B) grs1 and dom haven't been kept in sync well generally.. so they are often mor different than you may expect
11:46 * barton          waves at cait.
11:45 * cait            waves
11:45 ashimema          then I just diff the backup and the newly created ones.. and make sure I've not accidentally dropped anything from the backup that I didn't want to
11:45 barton            ok, I think I'll skip them...
11:45 ashimema          basically, I run through the step 2 (using scripts to create new dom configs) having taken a backups of the existing ones..
11:44 ashimema          barton: those additional steps are highly frowned upon by Jared.. you have been warned ;)
11:44 barton            I've been testing your process against 3.14.05 -- I don't think that my biblio-koha-indexdefs.xml ends up looking like yours...
11:43 ashimema          zebra is a complex beast
11:43 cait              ping gmcharlt
11:43 ashimema          yeah, that's true.
11:42 khall             it would also be nice to have a system to add new indexes to the advanced search pulldown, but that's a different ball of wax. We just do it with jquery at the moment
11:42 ashimema          mmm.
11:42 ashimema          along with the packages patch I've put together it'll allow proper customisation of zebra config all over I beleive
11:42 barton            your 'Additional steps for DOM indexing...' section -- you've got a couple of lines that get added to ' ~/kohaclone/etc/zebradb/marc_defs/marc21/biblios/biblio-koha-indexdefs.xml'
11:41 khall             that is exactly what we've discussed doing as well
11:41 ashimema          (I'll sign it off though.. ;) )
11:41 khall             np! ; )
11:41 barton            oh, he's used to it.
11:41 ashimema          haha.. sorry khall.. volunteering you ;)
11:40 khall             indeed!
11:40 ashimema          go ahead
11:40 ashimema          My approach would be to append an array (based upon a directive in koha-conf.xml) to the indexes array in Search.pm. sure khall could nock that up in minutes ;)
11:40 barton            oh, one more question.
11:40 barton            ... although frankly, I've got a pile of zebra index tickets that I'm going to be able to close real soon now...
11:39 ashimema          lol
11:39 barton            ;-)
11:39 barton            who does?
11:39 ashimema          but I don't really have the time to do it at the moment.
11:39 ashimema          should be sort of straight forward to add a syspref or config option to enable simple additon of indexes to for koha's array..
11:38 barton            well, thanks for the update... that will give larrby, khall and me something to talk aobut today :-)
11:37 barton            exactly.
11:37 ashimema          it's the last missing step really ;)
11:37 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=11910 enhancement, P5 - low, ---, gmcharlt, NEW , Adding a zebra index should be a configuration change.
11:37 ashimema          and of course.. your bug 11910
11:36 huginn            04Bug 12216: enhancement, P4, ---, martin.renvoize, Needs Signoff , One should be able to override zebra configuration on a per instance basis
11:36 ashimema          may be worth looking at: http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=12216
11:35 barton            right.
11:35 ashimema          gitified could be interesting.. as managing the zebra config will be a tad different in terms of all the locations..
11:34 ashimema          in fact, we're now building our own repositories and going the whole pacakge appraoch
11:34 barton            but this gives me somewhere to start.
11:34 ashimema          yeah, we've been doing that..
11:34 barton            ... but we're also moving toward a gitified package install.
11:33 ashimema          :)
11:33 ashimema          else your changes won't take effect
11:33 barton            right... I think that we do basically the same...
11:33 ashimema          the important bit is running through the makefile, make and make upgrade steps.. then rebuilding zebra.
11:33 ashimema          which is a dev install using git etc.. so you may need to alter some apths etc..
11:33 barton            nod.
11:32 barton            et voila!
11:32 ashimema          The command set in there is based on how we install koha though..
11:32 * barton          refreshes wiki/MRenvoize/zebra ...
11:31 barton            aha!
11:31 ashimema          ok barton, I've lifted the relevant page from our internal wiki now.
11:30 barton            Ironically, it's never actually notified me of anything, either.
11:29 barton            hense the dbus message above.
11:29 barton            I don't think that my irssi 'ring the bell when my name shows up' plugin likes unicode.
11:27 barton            heh.
11:26 huginn            barton: The current temperature in Kikut, Oslo, Norway is 6.9°C (1:16 PM CEST on May 09, 2014). Conditions: Clear. Humidity: 69%. Dew Point: 2.0°C. Windchill: 6.0°C. Pressure: 29.62 in 1003 hPa (Steady).
11:26 barton            @wunder oslo
11:25 magnuse           howdy barton
11:23 barton            morning magnuse :-)
11:22 barton            coffee++
11:22 magnuse           yay! :-)
11:20 * barton          has coffee now... Coffee: |==>|................| 20 % <== this means that I'm functional ;-)
11:14 barton            so... I've been working through your instructions for updating zebra indexes... you suggest working in kohaclone, then running the 'koha update procedure'... but the link to that is broken.
11:12 ashimema          just in a call.. will reply in a minute
11:11 barton            not enough coffee yet.
11:11 * barton          shakes his head...
11:10 ashimema          I'm here barton ;)
11:09 huginn            barton: The operation succeeded.
11:09 barton            @later tell ashimema there's a broken link on your 'adding a link to zebra' page: 'Updating Koha Install' does not exist on the koha wiki.
11:05 khall             mornin all!
11:05 barton            morning khall!
11:03 ashimema          hi
11:00 barton            ping ashimema
10:58 * magnuse         hums "it's oh so quiet" almost inaudibly
10:52 ashimema          super quiet
10:26 cait              hm friday
10:25 cait              quiet freiday
08:32 ashimema          :)
08:00 cait              ashimema: read your pm
07:57 cait              morning
07:57 ashimema          morning cait
07:54 cait              ping gmcharlt
07:48 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=12221 enhancement, P5 - low, ---, pasi.kallinen, NEW , Remove TT statement placeholders from translatable strings
07:48 paxed             magnuse: re. trimming placeholders from po-files, see bug 12221
07:06 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=9593 normal, P5 - low, ---, sophie.meynieux, Pushed to Master , Prices not imported correctly from a staged file
07:06 sophie_m          khall_away: I only receive now your notification on bug 9593, sorry for not answering before, but I see that now it is pushed
07:02 alex_a            bonjour
06:58 sophie_m          hello magnuse and #koha
06:58 reiveune          hi magnuse
06:58 magnuse           hiya sophie_m and reiveune
06:52 wahanui           bonjour, reiveune
06:52 reiveune          hello
06:42 * magnuse         waves
05:59 huginn            cait: The current temperature in Taegerwilen, Taegerwilen, Germany is 12.7°C (7:55 AM CEST on May 09, 2014). Conditions: Light Drizzle. Humidity: 90%. Dew Point: 11.0°C. Pressure: 30.06 in 1018 hPa (Steady).
05:59 cait              @wunder Konstanz
05:58 cait              h dcook
05:50 dcook             yo cait
05:07 cait              hi #koha
03:51 eythian           http://blog.petflow.com/truly-astonishing-i-had-no-idea-they-were-this-smart-must-see/ <-- honey badgers
03:45 eythian           Something I sent on wednesday appeared to only arrive this morning.
03:45 eythian           the koha list definitely seems to be playing up.
03:35 rangi             heh
03:34 dcook             Gotta make sure someone is around to keep rangi in working order :p
03:34 eythian           yeah, we talked about rangi :)
03:34 dcook             Ack!
03:34 wizzyrea          j/k
03:34 wizzyrea          I feel ok having a heart attack now.
03:34 eythian           (actually, I don't know that my old one expired, but it was getting close.)
03:33 * eythian         once again has a valid first aid qualification
03:28 * dcook           waves
03:27 eythian           hi
02:01 dcook             That link tracker is a nice touch
02:01 dcook             rangi++
01:22 * dcook           takes a cupcake break.
01:22 dcook             Well, maybe they all are.
01:21 dcook             Actually, I guess those first few posts would be an easy patch..
01:17 * dcook           is pretty sure he doesn't even sound coherent anymore
01:17 dcook             All the patches...
01:17 dcook             It groups and deduplicates bib level and item level links, so that you can get them in the results as well..
01:17 dcook             Now I'm remembering why we made a change :p
01:16 rangi             sweet
01:16 dcook             Sounds good. I'm working on our new Koha, so trying to combine all the changes..
01:15 rangi             and whatever conditions you want
01:15 rangi             just select * from linktracker
01:15 dcook             Sweet as. Thanks :)
01:15 rangi             nope
01:15 dcook             Cool. Nothing else involved?
01:14 rangi             yep
01:14 dcook             So you set the pref, and then you get the data using your own SQL queries from scratch?
01:14 rangi             yep
01:14 dcook             Actually, while you're around rangi, can you shed some light on the link tracking?
01:13 dcook             Thought I'd make folks aware in any case :p
01:12 dcook             I have a few re-writes to the XSLT that will come up eventually anyway..
01:12 * dcook           hasn't had breakfast so he might be a bit ornery
01:12 dcook             I think Owen was bang on about using the existing pref from the start
01:12 dcook             I don't really see the purpose of that syspref anyway.
01:11 dcook             Maybe to write the patch. Probably not the test plan though.
01:11 dcook             Hehe. I don't know about that.
01:10 rangi             you've ranted more than it would take to write a patch :)
01:09 dcook             night tcohen!
01:09 dcook             Mmm pizza
01:09 tcohen            night #koha, pizza time :-D
01:09 tcohen            networking issues
01:04 huginn            dcook: The operation succeeded.
01:04 dcook             @later tell tcohen Possibly, although we use a non-standard FS layout, so probably not for us. Seems like a worthwhile goal overall though!
01:03 dcook             Sorry for spamming everyone
01:03 * dcook           is having a mega rant on Bugzilla it seems
00:40 huginn            04Bug http://bugs.koha-community.org/bugzilla3/show_bug.cgi?id=11962 enhancement, P5 - low, ---, tomascohen, NEW , New 'cluster' install mode
00:40 tcohen            dcook, as a tar user, do u think you could benefit from bug 11962?
00:36 tcohen            there might be some glitches
00:36 tcohen            thanks!
00:36 rangi             yep ill have a look after lunch
00:36 tcohen            i've been focused on the packages
00:35 tcohen            rangi, if you have some time would you take a look at https://github.com/tomascohen/koha/tree/bug_11404_apache2.4 ?
00:30 * tcohen          hopes he finds the time to publish Memoize::Memcached v1.00 soon
00:29 rangi             its a very low barrier to entry, so you could even join the debian perl group and update the modules yourself
00:29 tcohen            thanks rangi
00:29 rangi             http://pkg-perl.alioth.debian.org/
00:29 wahanui           it has been said that the best thing to do is make the password VERY secure
00:29 rangi             the best thing to do
00:28 tcohen            btw, how do i ask Debian perl maintainers to update some lib I maintain?
00:28 rangi             since we want/need to package upstream perl modules sometimes too
00:28 tcohen            :-D
00:27 rangi             maybe i could do a packaging tutorial or something at kohacon
00:27 rangi             but it doesnt hurt to know about it
00:26 rangi             since we are upstream as well, we dont have to do that, we can edit the koha-conf.xml.in file ourselves etc
00:26 rangi             i would add the changes i need as a debian patch ..
00:25 rangi             say if i was packaging koha, but wasnt a developer of koha
00:25 rangi             when you arent the upstream
00:25 rangi             you usually do this
00:25 tcohen            heh
00:25 tcohen            damn rangi, more stuff to read before bed!
00:25 rangi             http://pkg-perl.alioth.debian.org/howto/quilt.html
00:24 rangi             i was fixing debian bugs last weekend with debian/patches for stuff, using quilt
00:24 tcohen            quilt?
00:23 rangi             you can do it with quilt quite easily
00:23 dcook             As a non-package user, I'm very interested
00:23 rangi             ahh right :)
00:23 rangi             so if we can do it, we should
00:23 tcohen            i just didn't know it before i read the ppa docs
00:23 * dcook           perks up
00:23 rangi             however that doesnt help the people using tar
00:23 rangi             yep
00:22 tcohen            i was just commenting that another approach (rather then supporting everything in the same code) is used too
00:22 rangi             if you are building for unstable, you should make it work with the unstable apache
00:22 tcohen            exactly
00:22 rangi             thats not specific to ppa
00:22 tcohen            so i don't need that
00:22 tcohen            i already have the patches for it to work with either version
00:22 rangi             yeah thats the same with any package
00:21 tcohen            so for the same codebase, one provides the patches for (for example) a specific apache version
00:21 rangi             only where needed
00:21 tcohen            i just noticed that the ppa's expect us to provide patches specific for each dist build
00:20 rangi             nope
00:20 dcook             Is this Debian's introduction of PPAs?
00:20 dcook             tcohen: Well, welcome home belatedly!
00:19 rangi             you have to wait for it to get built, and it can be in the queue for hours
00:19 rangi             its a lot faster to just build your own packages using the build-git-snapshot and pbuilder .. a lot faster
00:19 tcohen            i might set my own just for testing purposes
00:18 tcohen            rangi: i don't think the ppa's should substitute our repo, I was thinking more of the way they work and expect us to send our stuff
00:17 tcohen            it was a 28 day trip in europe
00:16 tcohen            yes, i got back two weeks ago
00:15 dcook             I only saw some photos but it looked interesting!
00:15 dcook             I was wondering if you were back from your honeymoon ;)
00:15 dcook             Sounds like an all right day
00:15 tcohen            work + meetings + tennis -> home
00:14 dcook             Back home?
00:14 tcohen            \o
00:13 dcook             o/ tcohen!
00:07 huginn            dcook: I'll give you the answer just as soon as RDA is ready
00:07 dcook             @$#%^#