Time Nick Message 09:44 kados morning owen 03:43 osmoze hello 22:36 chris that would be cool 22:36 thd :) 22:36 thd chris: Dublin Core actually scares me because it does not use code numbers and ignores the degree of detail found in full MARC. 22:33 thd chris: There may be a standards compliant way of identifying the lesser degree of record detail that Koha had originally used with Dublin Core. 22:21 thd chris: MANY ILS systems require knowing how to encode 650##$a$x$z$y to use them at all or get much advantage from them. Paul's work already goes a significant way to correct that. I like to have data views that support the value of MARC but guide the user with meaningful labels where something is not self-evident. No code numbers should trouble or frighten the user. 22:17 chris cool 22:17 thd chris: I am keen to make the system work well for the end user who should never need to know MARC. 22:16 chris cool 22:16 thd chris: My background is in bookselling, where I hid LIS stuff from the user while taking advantage of MARC records. 22:15 chris in making sure it works for non-MARC .. ill leave the MARC bits too the people who know more 22:15 chris im sure we will :) i was just saying that thats what my main focus with the zebra integration will be 22:14 thd chris: We will hide MARC from HLT. 22:14 chris as long as we dont turn into one of the big ILS vendors and forget about all the little ppl who made it possible 22:14 chris yep, as i said thats great 22:13 thd chris: The potential new class of customers, would be much larger and potentially a much better source of business than the instutuions for which Koha had previously offered a solution. 22:12 chris (one of the 11) 22:11 chris thd: indeed and thats great, as long as we dont lose HLT (the original client and reason Koha exists) 22:11 chris so just want to make sure we dont repeat our mistakes :) 22:10 chris its just that when marc was introduced, the non-marc cataloguing features lagged behind and werent tested as well as the marc ones (as much from my fault as anyone elses) 22:10 thd chris: However, proper MARC support implemented well should give you the possibilty of serving a whole new class of potential customers. 22:10 chris yep 22:08 thd chris: I would imagine, there will always be ways to hide MARC from the user who may be frightened by 650 $a$x. 22:07 chris then=the 22:06 chris yep, i dont imagine it will be a problem, as long as we make sure that then non-MARC cataloguing pages update the files zebra indexes 22:05 thd chris: If you read paul's migration email carefully, he seems to intend to keep the database structure along with blob storage or whatever for MARC records that zebra can index. 22:03 chris most of them are corporate or special interest libraries 22:02 chris thats the one 22:02 thd chris: Did I interpret that correctly? 22:02 thd chris: I saw your post about 11 of your 13 customers prefer to never see MARC. 21:49 chris am now 20:56 thd chris: are you stiil there? 18:56 kados thd: right 18:56 kados thd: there are other ways to tap into the api but a perl-specific interface would be ideal 18:56 thd kados: you mean it does not support all the features that the C or whatever API does? 18:55 kados thd: it was developed by a third party who's vanished ;-) 18:55 kados thd: the perl api for zebra is not fully developed 18:54 thd kados: what is missing form the api currently? 18:54 kados so that's something to keep in mind 18:53 chris cool 18:53 kados perl api that is 18:53 kados he's interested in expanding the zebra api 18:53 kados I spoke to sebastian this morning 18:53 thd chris: fortunately, we are following after a significant degree of format conversion to MARC21 and UNIMARC 18:53 chris cool 18:53 kados ok ... summary sent 18:52 chris yeah, thats not even worrying about the other 20 thousand marcs (chris might be exaggerating slightly) 18:51 thd chris: I was wrestling with magic tricks to overcome the MARC21 UNICODE divide. I have not found the right spell book :) 18:50 chris i think moving marc out of the db will be the best thing we can do 18:49 chris it gets worse when you try to support more than one flavour 18:48 chris yeah 18:47 thd There has been a discussion on the MARC listserv just recently about RDB and MARC with the usual negative conclusions. The addition of XML was considered to be no advantage for the databse design problem. 18:43 thd :) 18:42 chris thd: i didnt even attempt to get MARC in .. i left that for the more masochistic programmers to try :-) 18:42 chris :) 18:42 chris both = lots of 18:42 chris yep kados, so as long as we can offer both views, im happy 18:41 thd the article has the same conclusions koha developers have discovered for themselves with more effort 18:41 rach ah right 18:41 chris and the refugee woman was quite keen (student librarian from vic) 18:40 rach do they? mad buggers 18:40 chris unido 18:40 kados because if you're talking about 245a or 246b ;-) 18:40 chris :) 18:40 kados well it makes a difference ;-) 18:40 kados right 18:40 chris instead of title 18:40 kados which is why it's so nice 18:40 chris some ppl just love to see 245a 18:40 kados without changing the api at all 18:39 chris yeah 18:39 kados with zebra you can change the underlying record format 18:39 chris i think in the near future anyway 18:39 kados not necessarily 18:39 chris anyway, dont get me started on marc .. we are stuck with it :) 18:38 chris storing MARC isnt 18:38 chris storing bibliographical data in a relational db is easy 18:38 thd Green, Rebecca. Design of a relational database for large-scale bibliographic retrieval. Information technology and libraries v. 15, no. 4 (Dec. 1996) p. 207-221. 18:38 kados chris: wow ... I didn't know you had that many koha clients! 18:37 chris and 2 who do 18:37 chris i counted 11 clients of ours, who use koha and dont ever want to to see MARC 18:36 chris is integrating with zebra in such a way that we can still offer our 2 views of the data 18:36 chris so our main task as i see it 18:36 chris heh 18:35 mason esp. the pic of the zebra 18:35 mason looks nice 18:34 thd I am having difficulty getting a question through on either the MARC or AUTOCAT listserves. I never had trouble on the MARC list before a few months ago. I think my *.com address is being identified as a potential source of spam. Maybe I need to register a *.org address. : / 18:33 kados yep 18:33 chris and end up with a ton of redundant data 18:32 chris you lose all the advantages of a RDBMS 18:32 chris marc in sql makes little sense really, when i think about it 18:31 chris for the marc data that is 18:31 kados mason: we're going to be moving from MARC-in-SQL to using Zebra as the primary storage and retrieval engine for Koha 18:31 kados mason: there's a new Koha list: koha-zebra 18:30 kados mason: sorry ... writing a summary ... 18:28 chris http://indexdata.dk/zebra/ <-- that zebra 18:28 chris nope :) 18:26 mason A system with Zebra installed acts as a dedicated router. With Zebra, your machine exchanges routing information with other routers using routing protocols. Zebra uses this information to update the kernel routing table so that the right data goes to the right place. You can dynamically change the configuration and you may view routing table information from the Zebra terminal interface. 18:26 mason this zebra? 18:25 thd kados: I have not had much occasion to study UNIMARC though. I have been learning much more about UNIMARC in the past few days :) 18:24 thd kados: My MARC familiarity is reasonably good, although, I gained much of my experience with MARC outside a library setting. Do not let my confession about spending years as a bookseller mislead you. William Moen, leading research into MARC at UNT, also spent may years as a bookseller before adopting library science as a profession. That is usually done the other way round when librarians retire to take up bookselling. 18:23 mason tell be a bit about zebra 18:22 kados :-) 18:08 mason im just googling zebra 18:08 mason hi kados, my marc is thin and rusty ;) 18:03 kados mason: I'm putting together a summary of where we're at with Zebra integration for koha-zebra and I'm listing roles as well 18:03 kados mason: so ... is it ok if I list you as a MARC expert? ;-) 17:50 chris cool 17:49 kados composing a summary of our progress with zebra integration atm 17:47 chris heya kados 17:47 kados morning chris 17:40 chris morning 16:12 thd tim: It has helped me to think about the issues in migrating MARC records. Thankfully, I do not have to migrate a system immediately. 15:55 tim I hope it's helpful. 15:54 tim It's the first and only perl script I wrote except for a few test scripts. 15:24 thd tim: s/moring/moving 15:22 thd tim: I am not moving from Athena, but am just generally interested in the problems involved in moring MARC data between systems and how best to overcome them. 15:21 thd tim: I had seen your second post, which I believe was the one I had linked with the URL, where you corrected your code from your first message. 15:12 tim And I was also wondering if you're moving from Athena. 15:12 tim thd: did you see my post about the bug in the original marc coversion script? 15:11 tim Gonna try talking to the person I mean to talk to this time. 12:59 tim I did that early on and then took so long working on holdings. I need to check my work and see if it still makes sense to me. 12:58 tim To make the switch easier on staff. 12:58 tim If I remember right, most of the changes were to try to get it to display more like the Athena easy edit screens. 12:26 kados tim: (asside from holdings) 12:25 kados tim: did you have to redefine the marc links to match your records? 12:23 tim I seem to do that a lot. 12:23 hdl no problem :) 12:23 tim I'm talking to the wrong person :) 12:22 hdl I am french ;) 12:21 tim hdl: are you converting from Sagebrush Athena? 12:21 hdl no. 12:21 tim They would be in marc_tag_structure and marc_subfield_structure wouldn't they? 12:20 tim kados: I'm not sure how good I did at setting up the marc/db links, but could send them to you if you want them. 12:19 tim hdl: did you see my post about the bug in the original script? 12:19 kados tim: yes ... thanks very much ... I'm still waiting for the client to export his data 12:08 thd tim: I was most interested in the MARC record issue. I have no knowledge of circulation issues, they are not formally part of my problem space. I had been pleased to see the code that you had posted on the list for helping me think about the MARC records issue.