Time  Nick        Message
11:58 |Lupin|     nicomo: so one ccode per file format
11:57 |Lupin|     nicomo: the idea is that we will use ccodes to kep track of different file formats
11:57 nicomo      indeed
11:57 |Lupin|     nicomo: but at least if things do not work as expected once ccodes have been set up, I'll know that one possible cause of the problem is the nozebra indexing...
11:57 nicomo      Well : if you do, ping me, if not, don't worry :-)
11:56 |Lupin|     nicomo: I don't know whether I'll be able to determine this
11:56 nicomo      exactly
11:56 |Lupin|     nicomo: you want to know how well noZebra will be able to earch through C-Codes ?
11:55 |Lupin|     nicomo: it's just that I'm not sure I have understood what exactly interests you
11:55 |Lupin|     nicomo: sure
11:55 nicomo      keep me posted on this, will you?
11:55 nicomo      but I'd be interested in the result though
11:55 nicomo      again, I haven't tested this with noZebra
11:54 nicomo      you're welcome
11:54 |Lupin|     nicomo: thanks
11:54 |Lupin|     nicomo: I understand
11:54 nicomo      then it'll appear in koha2marc mappings
11:54 |Lupin|     aaaaaaah
11:53 nicomo      you have to create it in the marc framework first
11:53 |Lupin|     there is no possibility to select 995Z
11:53 nicomo      yes
11:53 |Lupin|     and in the 9xx one
11:53 |Lupin|     one for each marc block
11:53 |Lupin|     and then I have several drop downs
11:53 nicomo      yes
11:53 |Lupin|     then ccode
11:52 |Lupin|     I chose items
11:52 |Lupin|     in the koha to mark link
11:52 nicomo      my install show a textarea for this syspref
11:52 nicomo      which dropdown?
11:51 |Lupin|     nicomo: still one can select only those fields that are listed in the drop-down, as far as I can see
11:51 nicomo      but again : i never used noZebra, so you might want to check with someone else on this
11:50 nicomo      nozebra indexes are defined in the NoZebraIndexes syspref
11:49 nicomo      955$Z might be ok
11:49 nicomo      not really, except perhaps avoiding any subfield already used in the 955 recommandation
11:48 |Lupin|     nicomo: and, how do you tell the nozebra index where it is ?
11:48 |Lupin|     nicomo: any recommendaiton regarding the subfield ?
11:47 nicomo      you basically select the subfield you want to link ccodes to, then tell the index (either zebra of noZebra) where it's at
11:46 nicomo      not that I know of in unimarc, no
11:46 |Lupin|     nicomo: but when you install koha and ask for sample datas to be installed, it doesn't link ccode to any marc field either ?
11:45 |Lupin|     nicomo: we don't use zebra here since our collection is small
11:45 nicomo      then head to zebra's record.abs and tell it where you put ccode
11:45 nicomo      you have to pick a subfield : there's none prescribed
11:45 |Lupin|     nicomo: ok
11:44 nicomo      yes, just reading your question above
11:44 |Lupin|     nicomo: do you have acess to a unimarc set-up koha ?
11:43 nicomo      yes
11:43 |Lupin|     nicomo: here ?
11:36 |Lupin|     hi kf !
11:36 kf          hi Lupin
11:34 |Lupin|     hmm not sure about 995.8 actually, because this subfield is not listed in the Koha to marc section as a possible target for CCODE
11:30 |Lupin|     I think it's 995.8, but I'd appreciate a confirmation.
11:30 |Lupin|     Could someone with a Unimarc koha pleasee tell me to which MARC field the ccode is linked ?
11:29 |Lupin|     hi Amit
11:26 Amit        hi Lupin
11:24 |Lupin|     hello !
05:41 kf          good morning #koha
04:13 Amit        hi indradg
04:06 brendan     hi Jo
04:06 brendan     heya Amit
04:05 Jo          moring Amit
03:36 Amit        good morning #koha
03:36 Amit        hi chris, brendan, Jo
02:01 pianohacker 'night
22:36 chris       :)
22:36 rhcl        another great idea is dinner, and I'm ready! Later...
22:34 chris       sounds like a great idea
22:34 rhcl        'to add'
22:33 rhcl        When we finish our actual project, I'm thinking of setting up a 'kete' and trying to get the still living staff of the cancer hospital add their thoughts and memories to it.
22:32 rhcl        Very interesting.
22:31 rhcl        I can't find the link now, but I remember reading a complete page of the memories/writings of an early settler of some part of NZ--way back when. I think it was a transcribed diary.
22:31 chris       its a tool for building a community and a repository
22:31 chris       individuals/organisations
22:30 chris       thats what its for
22:30 rhcl        chris: still on? I have been looking at Kete over the past several weeks. From the examples on the site I see it would would be ideal for individuals to contribute to some collective goal.
22:20 Snow_Fox    true
22:20 gmcharlt    which would probably give it enough concurrency for what you want
22:20 Snow_Fox    and see if it works
22:19 Snow_Fox    ya i know what you mean, only way to see would be to test it
22:19 gmcharlt    Snow_Fox: I wouldn't count on that working - although since it uses sqlite for the borrowers database copy, it presumably wouldn't be too far from it using sqlite for the transactions record
22:18 Snow_Fox    it would simplfy everything
22:18 Snow_Fox    though
22:18 Snow_Fox    if we could use one file
22:18 rhcl        We are actually using the jumpbox version in a VM.  http://www.jumpbox.com/
22:18 Snow_Fox    so that we can go back and update all at once more or less
22:17 Snow_Fox    have the users create a file on a maped drive on one of our servers
22:17 Snow_Fox    is if the system goes down
22:17 Snow_Fox    the way i figure on setup
22:17 Snow_Fox    hrm
22:17 Sharon      one is the State gov't.  funny
22:17 rhcl        Yea, a lot of major universities use it. It's pretty polished and professionally well-done.
22:16 Sharon      there are at least 2 Kansas uses of Dspace
22:15 chris       never used it
22:15 chris       pass
22:14 Snow_Fox    im gonna assume that the offline circ system is not capible of writing of having multiple stations write to the same file on the network at the same time correct?
22:12 Sharon      cool!  I'll take a gander
22:11 rhcl        http://pastebin.com/d25764106
22:10 rhcl        We have about 25 books with 350-400 pages each, plus some odds and ends like newspaper articles.
22:08 Sharon      big? small? replicable?
22:07 rhcl        We are about halfway+ through a digitization project- Project Bloodroot
22:07 Sharon      awesome
22:06 chris       http://horowhenua.kete.net.nz/about
22:06 rhcl        I thought so.
22:06 Sharon      rhcl Kansas - part of the NExpress folks
22:06 chris       http://horowhenua.kete.net.nz/site/images/show/1683-foxton-senior-rugby-team-1959?private=false#comment-128
22:06 Sharon      local history stuff - pictures, family histories, etc.
22:06 rhcl        Sharon: where are you located?
22:05 chris       depending on what sort of project you are working on
22:05 rhcl        IMHO, it's worlds better than Greenstone.
22:05 chris       and check out kete too
22:05 rhcl        Look at DSpace.
22:05 Sharon      I'm suppose to be putting together a digitization pilot project, so I'm glad the linked to greenstone and others.  I need some educatin'
22:04 chris       :)
22:04 chris       course my director is also the president of the NZOSS
22:04 Sharon      we're doing an Open Source tech day workshop in August, so that's timely
22:04 chris       my director just twittered it
22:03 Sharon      I posted that on our Tech blog
22:01 rhcl        Yea, pretty good. I always forward articles like that to my director.
22:00 chris       its a nice balanced write up i thought
22:00 chris       well duraspace :)
21:59 chris       yep, dspace == fedora now
21:59 rhcl        I see DSpace got a mention.
21:52 chris       http://www.linux.com/news/enterprise/case-studies/26661-check-out-open-source-library-and-collections-management-softwar
21:22 pianohacker Bye
21:22 joetho      GOODBYE CRUEL WORLD
21:22 joetho      thanks for the tips,
21:22 joetho      I will peck at this sql crap later.
21:13 chris       ahh :-) hope you brought a few pair of shoes
21:13 gmcharlt    setting up booth and preparing to wear down our feet
21:12 chris       whats on the cards for tomorrow?
21:10 chris       ahh i hate when they do that
21:09 gmcharlt    annoying, that
21:09 gmcharlt    chris: nice enough; but given that it's a major metro hotel, gouging me for internet access
21:08 chris       hotel too
21:08 chris       gmcharlt: nice hote?
21:08 ricardo     Bye (out!)
21:08 ricardo     Thanks again
21:08 ricardo     chris: LOL! No problem  :)
21:08 chris       (sorry got distracted by my boss)
21:07 chris       good night :)
21:07 chris       and use the upload borrowers tool in 3
21:07 chris       you could do the select into outfile and then munge the file
21:07 ricardo     Well, it's past 10 PM here, and I'm still at work and haven't had dinner. Going home now. Take care everyone! And thank you Chris!  :)
21:06 ricardo     chris: I'll think about it. Thanks!  :)
21:04 ricardo     (records -> *biblio* records, I mean)
21:04 ricardo     chris: And I would then delete the records and do "bulkmarcimport"s afterwards?
21:03 chris       well, you;d just ignore that (ie you wouldnt use the upgraded one)
21:03 ricardo     chris: Yeah... That's the other option. I don't know what would happen to the biblio stuff then, though
21:03 chris       you can do the select into outfile
21:03 chris       OR
21:03 chris       then you can mysqldump and load it in
21:03 chris       so that your borrowers table at least gets upgraded
21:02 chris       or you can do the upgrade, paying ignoring the bibliographic stuff
21:01 ricardo     chris: OK. Thank you very much for the tip  :)
21:01 chris       load data infile '/tmp/borrowers' into borrowers (column_name,column_name.....)
21:00 ricardo     chris: The upgrade possible doesn't seem possible, I'm afraid (character set problems, and fields that were wrongly filled in our Koha 2.2.9 - like the "sensitive" 100a field in UNIMARC - and that now are (correctly) validated in Koha 3.0.x
21:00 chris       the next best way, is to do select * from borrowers into outfile '/tmp/borrowersdata';
20:59 chris       and not lose any historical stuff
20:59 chris       and then you have tables that you can then mysqldump and load straight into your koha 3
20:59 chris       then you end up with a db, that will work with then you can upgrade it to 3.0.3 etc
20:58 chris       and put it somewhere then follow the upgrade to 3.0.0 rules
20:58 chris       the best way to do it, is to take a copy of your 2.2.9 database
20:58 ricardo     chris: Meaning...?
20:58 chris       they dont have to
20:57 ricardo     chris: Oh, you're right. But those will also change :(
20:57 chris       itemnumbers is what it cares about
20:57 ricardo     chris: Ideally, those too. But for now, just borrowers (because biblionumbers will change and circulation historical data will probably get "confused" by that)
20:56 chris       the short answer is going to be, there is no tool in 2.2.9 to do it
20:56 chris       circulation data and accounts data too? or just borrowers?
20:54 ricardo     How do I export patron data in Koha 2.2.9 (so I can later import it in Koha 3.0.3)?
20:54 pianohacker chris: hey
20:54 ricardo     Wb chris !  :)
20:54 chris       back
20:51 joetho      SELECT  from items LEFT join authorised_values on (ccode = authorised_value) where lib IS null
20:48 joetho      Can't use an undefined value as an ARRAY reference at /home/sek/kohaclone/C4/Reports/Guided.pm line 412.
20:46 pianohacker joetho: What particular syntax error is it giving you? The fact that it worked at all might be a peculiarity of my setup
20:45 pianohacker Learn Latin American Spanish, then you won't have to care about the difference
20:45 ricardo     jdavidb: LOL
20:44 joetho      Yez, I zee
20:44 jwagner     joetho, the answer is Yes
20:44 pianohacker s in the table, z in the interface
20:44 joetho      s, right?
20:44 joetho      is authorised spelled with a z or an s?
20:43 jdavidb     atz's clever patch will also eliminate your duplicate bib issues, see?
20:43 jdavidb     Awesome!  Thanks, atz!
20:42 atz         DROP TABLE biblio;  -- done
20:40 joetho      In the interest of accuracy, I propose that I use only words and numbers that contain neither a zero nor "letter O".
20:39 joetho      but I don't see how that could happen with ccodes. One wrong character when you are importing huge batches should give thousands of errors, not 4 or 5 a week.
20:39 ricardo     gmcharlt: I think it's more like "Oh-0h"  (pun *fully* intended  ;-)
20:37 joetho      zeroes*
20:37 gmcharlt    uh-0h
20:37 joetho      An Un-named Employee left me a few code presents in the form of transposed ohs and seroes, but once you find them it's no big deal.
20:37 wizzyrea    even though I know it's a drum stick
20:37 wizzyrea    I read drumstick and I thought "chicken drumstick?"
20:37 pianohacker Accidental delete of ccodes by someone, odd imported data, who knows
20:36 joetho      but assuming I DO find some weird values in there- how did they get there?
20:36 pianohacker I'm starting to talk in soundbites, God help me
20:36 munin       jdavidb: Quote #9: "pianohacker ponders drumstick->ear as a method of food acquisition...We haven't gone to this good mexican restaurant in a while..." (added by wizzyrea at 08:23 PM, June 19, 2009)
20:36 jdavidb     @quote random
20:36 ricardo     Hi richard
20:36 joetho      syntax. I'm pecking at it.
20:35 pianohacker Hrmm. Syntax error, or just don't pull nothin' up?
20:35 joetho      pianoist: I haven't got it to work yet. Sqlirrelly something or other in there.
20:35 richard     hi
20:34 ricardo     pianohacker: Amusingly and sadly insightful!
20:34 munin       wizzyrea: Quote #10: "< pianohacker> You helped start an open source project; clearly your sense of what to avoid to make your life easier has been impaired for a while :)" (added by chris at 07:59 PM, June 23, 2009)
20:34 wizzyrea    @quote random
20:33 munin       pianohacker: Quote #5: "<jwagner> Why is it every Koha rock I turn over produces a zillion (metaphorical) ants, each with a new question????" (added by kf at 01:13 PM, June 12, 2009)
20:33 pianohacker @quote random
20:33 joetho      I resolved all my actual "null" itypes etc, but not the ones that are THERE but no authorized.
20:33 Sharon      cool beans
20:33 ricardo     RDA?
20:32 munin       ricardo: I'll give you the answer as soon as RDA is ready
20:32 ricardo     @quote
20:32 munin       ricardo: I suck
20:32 ricardo     @qote
20:32 pianohacker joetho, Sharon: The SQL I posted above should find those, ah, "dysfunctional orphans", rather than just null ccodes
20:32 ricardo     jdavidb: eheh
20:31 munin       wizzyrea: The operation succeeded.  Quote #14 added.
20:31 wizzyrea    @quote add joetho: I only looked for missing ccodes- not dysfunctional orphans.
20:31 Sharon      joetho I have sql reports to find null itypes and ccodes and locations.
20:31 munin       gmcharlt: The operation succeeded.  Quote #13 added.
20:31 gmcharlt    @quote add <atz> ricardo: ask and ye shall receive.... eventually, ye shall stop asking.
20:31 jdavidb     Bugzilla?  Naah...front page of the Wiki.  That's priceless!
20:31 ricardo     atz: LOL! I think that would be a good "quip" for Koha's Bugzilla  ;-)
20:30 atz         ricardo: ask and ye shall receive.... eventually, ye shall stop asking.
20:29 pianohacker O_o There is no doubt a section for exactly that, this being the internet
20:29 joetho      I find those on match.com
20:29 joetho      I only looked for missing ccodes- not dysfunctional orphans.
20:29 pianohacker I think you have gmcharlt to blame for that
20:29 pianohacker Heh
20:28 ricardo     pianohacker: OK. I found the screen for adding a matching rule... but I'm beginning to think that I wish I *didn't*!  ;-)
20:28 pianohacker Should pull up anything with a broken ccode
20:28 pianohacker joetho: SELECT * from items left join authorised_values on (ccode = authorised_value) where lib is null;
20:28 joetho      it definitely includes expanded functionality for defining matching rules.
20:28 joetho      our development spec for this seems fairly extensive.
20:27 ricardo     pianohacker: OK. Thanks for the info. Let me check that
20:27 pianohacker ricardo: You first have to define record matching rules
20:27 joetho      I have been very careful about deleting authorized values.
20:27 ricardo     Is it because that development work is being done in this area, that the only "Record matching rule" (available in its dropdown list) is "Do not look matching records" (in the "Stage MARC Records for Import" screen)?
20:27 joetho      hmmm.
20:26 pianohacker joetho: Is it possible that some of your records have ccodes that are no longer defined in the authorised_values table ?
20:25 ricardo     On a related note...
20:25 atz         joetho: you didn't find NULL values in the records?
20:24 jdavidb     Awwwwww...we wuvs you, wizzyrea!
20:24 joetho      well..... bucks up.
20:23 gmcharlt    c'mon, wizzyrea - the whole point of social apps is that you're obligated to keep up with *everything*
20:23 pianohacker Not everyone is glued to their email inboxes (not entirely a bad thing...)
20:23 joetho      btw, @atz suggestion for my earlier "null" question didn't work.
20:23 wizzyrea    *sigh*
20:23 ricardo     Poor random!  ;-)
20:23 munin       ricardo: Karma for "random" has been increased 0 times and decreased 1 time for a total karma of -1.
20:23 ricardo     @karma random
20:23 jdavidb     Yah, wizzyrea...if you kept up with the lists, whilst cutting up on here, you'd have known what pianohacker was talking about...c'mon...do try to keep up.  :P
20:22 ricardo     joetho: Really? That's great! That means I can stop working on this data migration and wait for you to finish the specs / development ;-)
20:22 pianohacker random--
20:22 pianohacker Hrmph. Mailing list
20:21 wizzyrea    wow, that was random lol (pianohacker)
20:21 pianohacker If you're not careful, you'll be assigned cleanup tasks for the rest of your career
20:21 joetho      thus our discussion of the specs for bib merging development.
20:21 pianohacker atz++
20:21 wizzyrea    gmcharlt: you crack me up. :)
20:20 wizzyrea    ooh, that's the question of the day
20:20 ricardo     Question: the easy way to remove duplicate bibliographic records in Koha (for a given item type) is...?
20:19 gmcharlt    ooh, shiny!
20:19 joetho      how about ten percent of the money spent on it?
20:19 ricardo     gmcharlt: eheh
20:19 gmcharlt    please send me at least 10% of any money wagered on that question ;)
20:18 ricardo     wizzyrea: LOL!
20:18 wizzyrea    pbbbt
20:18 joetho      see I toldja
20:18 gmcharlt    wizzyrea: yes
20:17 ricardo     joetho: LOL
20:17 joetho      uh oh dads here
20:17 pianohacker joetho: LibLime specs I've seen in the past were creative commons licensed
20:16 wizzyrea    oh for pity's sake. Yes, look closely at what you have, it's printed at the bottom
20:16 joetho      ha. a hesitation.
20:16 joetho      yer totally sure.
20:16 wizzyrea    yes
20:16 joetho      yer sure.
20:16 wizzyrea    re: joetho
20:16 wizzyrea    no, they are creative commons licensed
20:16 joetho      not sure if they are vendor-specific-proprietary at this point. A very murky subject, that.
20:14 jwagner     wizzyrea, Sorry, I've been off on another system.  No, I haven't seen the NEKLS specs -- can you point me to them?
20:14 joetho      I think this will go a long way toward solving problems with batch deletions
20:14 joetho      we are also spec-ing out a development to deal with merging bibs together, that preserves all the item data.
20:06 wizzyrea    (and I'm sure we'd be interested in seeing what you're doing for your client :P)
20:06 wizzyrea    jwagner: have you looked at the NEKLS specs for bib maintenance?
19:40 ricardo     chris: Bye Chris... and thanks! :)
19:36 jwagner     Happy commuting, chris!
19:36 chris       right off to ride my bus
19:36 ricardo     jwagner: Eheh
19:36 jwagner     Nope.  I didn't even say "yesterday" !
19:34 ricardo     (just kidding...)
19:33 ricardo     jwagner: Hi Jane! That's great! Thanks for the info  :)  And don't worry, I won't "nail" you to a deadline... But you said "tomorrow", right?  ;-)
19:32 jwagner     PTFS is working on batch item edit for a client -- code is still in test, so I'm not sure when it will be available.
19:32 ricardo     (I must say I'm not fond of data migrations, to say the least...)
19:30 ricardo     chris: OK. Thanks for that info, too
19:30 chris       but im not sure where its at
19:30 chris       im fairly sure there is work being done on bulkedit/deletes of records
19:29 chris       righto
19:29 ricardo     Yeah... In my case "rebuild_nozebra"
19:29 chris       -k
19:29 chris       then rebuild_zebra.plk
19:29 ricardo     chris: *nod*
19:29 chris       BACKUP THE DB FIRST :-)
19:29 ricardo     chris: Thank you very much for the query.  :D
19:29 ricardo     chris: Right.  I must admit that I don't feel very comfortable with editing directly the tables... I wish there was some kind of API for this (or an option for bulkmarcimport). Oh well
19:28 chris       delete from biblio,biblioitems,items where biblioitems.biblionumber = biblio.biblionumber and items.biblioitemnumber = biblioitems.biblioitemnumber and biblioitems.itemtype='SOMETHING'
19:27 chris       if it was at itemlevel, you would have to edit the marc (removing just the items of the certain type)
19:27 ricardo     chris: Yeah. Thank God for small favours, right?  ;-)
19:27 chris       *nod*
19:27 ricardo     chris: I mean that I set the "item-level_itypes" System Preference to OFF
19:27 chris       ricardo: cool, in that case its easier, if it was at item level you would definitely have to script it
19:26 Snow_Fox    roger that thanks chris
19:26 chris       koha-devel mailing list is your best bet Snow_Fox
19:26 ricardo     chris: Right. But, in my case, i have item types NOT set at "item level"
19:26 munin       chris: I have not seen kyle.
19:26 chris       @seen kyle
19:26 Snow_Fox    heh
19:26 Snow_Fox    duh no x3
19:26 Snow_Fox    .seen kyle
19:26 chris       itemtype can be at itemlevel
19:25 chris       the trick is
19:25 chris       biblioitems contains the marc and marcxml blobs
19:25 chris       you want to get rid of biblio, biblioitems and items;
19:24 ricardo     chris: That's what I'm afraid of  :-S  And afraid because, besides the biblio.* tables, I don't know if I have to delete other things because of the MARC representations
19:24 chris       my guess is yes, but kyle is who could answer definitively Snow_Fox
19:24 Snow_Fox    is that refering to a item hold?
19:24 Snow_Fox    hey on the offline circ program when it says that it wont record a "fills a request"
19:23 chris       ricardo: i think you would have to do it in the db with some sql
19:23 ricardo     Eheh
19:23 chris       lol
19:23 ricardo     chris: I thought that Koha related work made you a millionaire and that you hadn't to work anymore  ;-)
19:22 chris       7.21am here, just feeding my son breakfast then i have to catch a bus to work :)
19:22 ricardo     Quick question: is there a way in Koha to delete all bibliographic records that belong to a specific "Item Type"?
19:21 ricardo     (it's 20H21 here in Portugal  ;-)
19:21 ricardo     "Morning" chris!  :)
19:20 chris       morning
19:18 mib_s1ex4e  cool, thankx
19:18 ebegin      mib_s1ex4e, this is for anybody who want to talk about koha.  both librarian and coders
19:17 mib_s1ex4e  is this just for koha writers or for koha cataloging?  Need to know if I need to be in here or not..lol
19:11 atz         or whatever your field of choice is
19:10 atz         select itemnumber from items where ccode IS NULL;
19:10 atz         NULL is the value for no value.
19:10 atz         joetho: right
18:35 joetho      items with no collection code or something like that? Nope. I am perplexed and stymied.
18:34 joetho      what are these -null- things in circ reports??
17:03 |Lupin|     bye "koha !
16:32 ebegin      paul_p, good.  Thanks for the info.
16:30 paul_p      so, it's not as nice & fancy as with jquery "active", but it still works
16:30 paul_p      BUT JQuery has a fallback++ without jquery
16:29 paul_p      ebegin: lynx don't support javascript.
16:00 |Lupin|     ebegin: no, I don't think so
15:58 ebegin      |Lupin|, does lynx support JQuery?
15:24 |Lupin|     it may indeed very well be that I didn't instal the example values...
15:22 |Lupin|     kf & ebegin ok, thanks !
15:22 Snow_Fox    with the offline transaction tool it notes that the following items will not be process or checked out, and was wonder what it ment by Item barcode fills a request
15:21 ebegin      |Lupin|, IIRC, the CCODE are part of the optional data that is added during the websetup.
15:19 kf          I dont know why, perhaps you chose not to install example values during isntallation, or its a UNIMARC / french isntallation thing?
15:19 kf          just use CCODE as category code
15:19 |Lupin|     kf: just surprised that it is not there already
15:19 |Lupin|     kf: oka !
15:16 kf          |Lupin|: when adding a new category you will add the first value in the same step
15:11 |Lupin|     kf: ok
15:09 kf          |Lupin|: try adding it with "new category"
15:09 |Lupin|     and then COUNTRY
15:09 |Lupin|     after BSort2 I have CAND
15:08 |Lupin|     kf: I just had a look to the manual. It mentions the CCODE category, but it does not appear in the drop-down I can see with categories
15:08 kf          |Lupin|: when there is no CCODE you can add it
15:08 kf          |Lupin|: you can import some predefined CCODEs during installation
15:07 kf          |Lupin|: its the right place
14:47 ebegin      paul_p, you think you can give a try to fill some of those values? http://wiki.koha.org/doku.php?id=en:development:hard_coded
14:29 |Lupin|     wet to administration -> authorized values and couldn't see anything
14:29 |Lupin|     can anybody help with ccodes pls ?
14:26 paul_p      ebegin: yep, end of 1999. but this one may not be that old. (which script ?)
14:23 ebegin      when was the first line of code written for Koha? 1999?
14:20 ebegin      no, i didn't :)
14:18 paul_p      ebegin: have you submitted a patch as of today ? If no, then you may be the 96 or 97th commiter !
14:18 |Lupin|     Is the ccode category predefined in Koha or is there something to do to activate it ?
14:17 ebegin      grr... thanks paul_p
14:17 paul_p      ebegin: no, afaik, it's hardcoded (a koha 1.0 piece of code...)
14:17 ebegin      good monring Snow_Fox
14:16 ebegin      are the accounttype (Rep, L, ...) are defined somewhere?
14:16 Snow_Fox    morning
13:58 |Lupin|     hey ebegin
13:45 ebegin      hey |Lupin|
13:42 |Lupin|     hello