Time Nick Message 10:59 kados yes I am speaking of two things 10:59 paul yes. 10:59 paul something like the ->dbh in Context iiuc 10:58 kados and passing it options, operation, etc.? 10:58 kados do you like my idea for having one ZOOM::Package method in Koha 10:57 kados ahh :-) 10:57 paul it's mike that warned me ;-) 10:57 paul yep. does nothing. 10:57 kados did you notify Mike Taylor? 10:57 kados hmmm, dropdb is buggy eh? 10:57 kados heh 10:56 paul mail sent to koha-devel, should arrive in 4 hours ;-) 10:56 paul (createdb is already in some code I wrote. note also that there is a dropdb... that is buggy and do nothing) 10:55 kados I may work on this routine a bit 10:54 kados and then we need an 'operation' for the $p->send method 10:54 kados it looks like we could pass in a hash object for $options 10:51 kados if we had a generic ZOOM::Package wrapper we could pass in options that we want 10:51 kados one of them allows 'createdb' 10:51 kados there are several methods in the ZOOM::Package module 10:50 kados I'm thinking we should change the name 10:48 kados ok 10:47 paul (not sure the commited one is the last I have) 10:47 paul I did 10:47 kados paul: did you write the zebra_create method in Biblio.pm? or did chris? 10:37 kados woo hoo! 10:36 paul (writing a mail on koha-devel to give -quite good- news from Ouest-provence) 10:35 paul maybe you're right, but with mod_perl that will be automatic behaviour I think (without destroy), while without mod_perl, we would have a destroy when the process ends 10:32 kados does that make sense? 10:32 kados $Zconn->destroy(); 10:32 kados update many records (all in fact) 10:32 kados $Zconn = new ZOOM::Connection(C4::Context->config("zebradb")); 10:32 kados I was hoping it would be faster to do: 10:32 kados $Zconn->destroy(); 10:31 kados update a record 10:31 kados $Zconn = new ZOOM::Connection(C4::Context->config("zebradb")); 10:31 kados we do: 10:31 kados right now 10:31 paul ??? not sure to understand what you mean. 10:30 kados http://search.cpan.org/~mirk/Net-Z3950-ZOOM-1.01/lib/ZOOM.pod#___top 10:30 kados paul: in your opinion, should we use the create()/connect() method to maintain the same object while updating records? 10:18 kados just some things I've noticed :-) 10:18 kados so at the beginning, it was about 10-20 per second ... but after 30,000 it was about 1 per second 10:18 kados i also noticed that it gets slower as more items are added 10:17 kados might be a speed improvement? 10:17 kados I wonder if we can have just one connection for the entire process 10:17 kados import is quite slow the connect - import one record - disconnect way 10:16 kados and zebra was getting overloaded with too many connections 10:16 kados but I think that was because we wern't destroying the zebra connection 10:16 kados the import would crash 10:15 kados I was seeing similar errors after importing about 200 items 10:15 kados XML that is 10:15 kados about invalid ZXML 10:15 kados as the last record that loaded threw an error 10:15 kados I'm still not sure all the records made it in 10:14 paul works nicely 10:14 kados try 'book not daniel' 10:14 paul (don't ask me what I entered as search term ;-) ) 10:13 paul The Book of Isaiah / 10:13 paul The Book of Daniel / 10:13 kados yep :-) 10:13 paul The Book of Ezekiel / 10:13 paul christian library it seems. 10:13 kados hi paul :-) 10:13 kados and chris's CQL query is working nicely 10:12 kados import of one of my client's data seems to have finished 10:12 paul paul greets kados 10:12 kados paul: http://opactest.liblime.com/cgi-bin/koha/search-test.pl? 23:00 thd kados: The bean counters who do not know MARC from Dublin Core and make all the decisions or the cataloguers who dream MARC and have no influence? 22:58 kados potential clients of course 22:57 thd kados: "For whom are 'fact sheets' intended? 22:56 kados that's fine 22:56 kados ok 22:53 thd kados: The ISBD mapping took about ten hours and I never audited it when I was done but I had to follow some ambiguous ISBD rules. At least there was no consequence for data loss in case of error there. 22:50 thd kados: I think it will be easier for me to change a default MARC 21 framework on my own system and send you an SQL file. I will have better access to the data than if I work on yours. 22:47 thd ? 22:46 thd kados: Who is the intended audience for Koha 'fact sheets'. 22:45 thd kados: Currently the migration path would be tedious except that hdl designed holdings for libraries that usually had simple reliable publication patterns. 22:43 thd kados: Record creation and modification is of comparable complexity to migration. 22:42 kados true 22:42 thd kados: Half the problem with Koha serials holdings now is the user interface requires careful application of machine encoded values, and is fairly brittle. 22:39 thd kados: Of course, you need a two way map in order not to be an evil trap for your customer like so many proprietary systems are. 22:38 thd kados: The better your support for mapping MARC holdings onto whatever the lower your costs will be. 22:37 kados of course :-) 22:36 thd kados: After all the nonsense to obtain a client, I assume that data migration is the next most time consuming part of your services. Am I correct? 22:33 thd kados: There is a wide variation in actual practise. If you can parse textual holdings fields even partly with regular expressions you will have a migration advantage. 22:31 thd kados: The most high demand holdings would have been redone as new records were created and systems supported their creation. 22:30 thd kados: Koha uses machine readable encodings for serials now, most serials holdings data was actually meant for human readability because the machine parsable holdings fields had not been designed. 22:28 thd kados: A large problem will be that accessible records will make extensive use of human readable relation encoding that some system needs to parse in order to work well. 22:26 thd kados: Fast is merely relative but part of my point is that the records may have tortured relations with one way mappings. Koha could find the implied relationships for easier access. 22:24 thd kados: You need to have access to the most troublesome records for testing that challenge all aspects of a system. 22:23 thd kados: Koha has historically used the simplest set of records as examples. 22:23 kados a well designed heirarchy can be very fast 22:23 thd kados: Obtaining bibliographic records is easy, finding complex holdings records may be more difficult.. 22:20 thd kados: I mean that only relatively as compared to flat data. Extra CPU cycles add up in large collections. 22:20 kados for real serials 22:20 kados like actual serial records 22:20 kados very specific examples 22:20 kados what we need are some specific case examples 22:20 kados I've been doing some work with them for PINES 22:19 kados very even 22:19 kados thd: nested sets are ver efficient 22:19 thd kados: Also, searching hierarchies can be inefficient. 22:19 kados you can traverse the heirarchy fairly easily 22:18 kados actually, with nested sets the search path is quite simple 22:18 thd kados: the difficulty is bringing a complex search path together when the records were not created in a neat top down manner. 22:18 kados if we used nested sets 22:18 kados peer nodes that is 22:18 kados nodes could be ordered by date 22:18 kados you could assign ids for each node in the tree 22:17 thd kados: Separate bibliographic records for various parts allow the inclusion of details about parts that would not have a proper place in a general record. 22:17 kados we could tackle an indefinite heirarchy using either an adjacency list or a nested set 22:17 kados that's not too hard to do 22:16 thd kados: Either way the system has to bring all the relations together. 22:15 thd kados: separate records allow for easier management, however, there could be one hyper-complex record managing it all. 22:14 thd kados: furthermore, the books may be monograph records and the back pocket supplements may be serial records. 22:14 kados what kinds of things can you do with the separate holdings records? 22:13 kados meaning what exactly ... you can interact with the records at each level in the tree? 22:12 thd kados: And each one can have separate holdings. 22:12 kados and there are links between the different records? is that what you're getting at? 22:12 thd kados: The criminal code may have its own index record linked to the other volumes. 22:11 thd kados: 22:11 thd kados: The criminal code part has a record for its several volumes both bibliographic and holdings. 22:10 thd kados: The set itself has holdings record for the set as whole. 22:08 thd kados: You have the annotated edition of the state laws with many volumes and a monograph record for the set. 22:07 kados I need some examples 22:07 kados that's a little too abstract for me 22:05 thd kados: holdings records can also be some arbitrary level below the title level. 22:04 kados I dont' quite follow 22:04 kados one serial record you mean? 22:04 kados ??? 22:03 thd kados: also, there is poor linking between bibliographic records where one record might cover the title as a whole, another an issue, another an article. 22:03 kados so let's focus on one thing at a time 22:03 kados that's true 22:02 thd kados: Koha support for serials is based on regular publication patterns that are predetermined. It is not very flexible for accommodating changes in publication patterns in the midst of a period or changing and irregular patterns. 21:59 kados could you explain it? 21:59 kados I don't really understand what that entails 21:59 thd kados: The issue is not for some support for supplements but supporting the complex part/whole relationships for supplements, special issues, etc. 21:57 kados thd: I've seen them in serials 21:57 kados thd: Koha supports summpements right now? 21:56 kados for two days of programming :-) 21:56 kados thd: probably not 100% CQL yet but it does a pretty good job 21:56 kados thd: that's the CQL test that chris built 21:56 kados thd: before I forget: http://opactest.liblime.com/cgi-bin/koha/search-test.pl 21:56 thd kados: Not that they are replaced when stolen at the poor libraries. 21:55 thd kados: Supplements are theft targets and vital to track for law libraries. 21:54 thd a supplement with the latest text revisions and case law updates. 21:53 thd kads: 5B was an arbitrary example for something like a serial back cover pocket insert that goes inside a volume of a law book. 21:51 kados right 21:50 thd kados: The big secrets are keeping the information about electronic resources current when your vendor is constantly changing the coverage without overwhel;ming the system maintenance staff. 21:49 thd kados: There are published papers on link resolvers. 21:48 thd kados: Karen Coyle did a fine job rewriting the documentation. 21:47 thd kados: I understand how OpenURL works. 21:47 thd kados: I have seen examples where the catalogues point to multiple hard copy and not yet come back next month when the embargo period is over electronic databases. 21:46 kados thd: isn't that what we're talking about providing? 21:45 thd kados: OpenURL yes but OpenURL needs detailed holdings for the resolver to use. 21:45 kados thd: ie, do you understand how they work? 21:45 kados thd: have you seen any really good examples of serials managment functionality in a proprietary system? 21:44 kados thd: about the gaps, spans of time, etc. 21:44 kados thd: we just need openurl 21:44 kados thd: linking is easy 21:44 thd kados: Holdings can link to electronic databases, or hard copy, electronic databases can link back. 21:43 thd kados: vendors will have gaps for single issues, spans of time, exclusion and coverage that changes on a sliding basis over time. 21:43 kados does it say _how_ to index? 21:43 kados I don't quite understand what you mean by that 21:42 kados and it can store information 'about indexing and abstracting database coverage'? 21:42 kados k ... holdings can store info about where full text databases are 21:42 thd kados: holdings can store information about indexing and abstracting database coverage as well as full text databases. 21:41 kados what's supplement 5B? 21:41 kados coudl you expand on that last bit? 21:40 thd kados: manage overlapping, embargoed, and gaps in electronic database coverage. 21:40 kados already does missing issues 21:40 kados links are cake too 21:39 kados with authorities 21:39 kados the title change should be easy 21:39 thd kados: The system has to manage the relationships between serial supplement 5B and a monograph after a title change. Display whole runs with links to the records and missing issues. Etc. 21:37 kados thd: what do the hierarchal relationships allow the system to do? 21:36 kados thd: what kinds of searching would we need to do? 21:35 thd kados: The hierarchal relations are the most troublesome for encoding, decoding, and searching. 21:33 thd kados: Unfortunately, the international standards are insufficiently detailed. They exist as a mere starting point constructed after the fact. 21:33 kados I _have_ afterall been drinking :-) 21:32 kados I need to digest that sentence 21:31 thd kados: there are also part, whole, etc. record relationships for linking between sets of holdings data and bibliographic records. 21:29 thd kados: yet what Koha can do already is as difficult to use as MARC and does not interface well with items although I know hdl partly fixed that. 21:28 kados interesting 21:28 thd kados: There is are NISO and ISO standards for serials. 21:28 kados thd: Kokha can do much of that already 21:27 thd kados: The problems are worst when the information has been recorded as a machine readability aid. 21:25 thd kados: complex publication patterns, special issues, supplements, missing issues, title changes of the same publication. 21:24 kados thd: can you explain what you mean by 'complexity'? 21:24 thd kados: You need a means of addressing the same complexity as MARC can. 21:23 kados thd: are you up to writing some on what MARC holdings (serial and otherwise) allow a library to do? 21:23 kados thd: exactly 21:20 thd kados: I see more clearly, searching complex MARC serials holdings is very inefficient. 21:20 kados where I do my best thinking :-) 21:20 chris cool :) 21:20 kados hehe yea, casa actually :-) 21:20 chris you using wifi at a bar? 21:18 kados then mapping it to MARC for those that want it 21:18 chris yep 21:18 kados then, it's a simple matter of doing it better inside Koha 21:18 chris ohh good point kados 21:18 thd kados: I have a solution to some of that for 2.X. 21:17 kados so what we need, is a list of things that the MARC standard for holdings can _DO_ 21:17 chris this is what i was saying about biblioitems .. the point of them was missed 21:17 kados we should also allow itemtypes at the item level 21:17 chris yep 21:17 kados for instance, we should be able to place reserves on specific items 21:16 kados that much is clear 21:16 kados we need to improve Koha's internal methods for managing items 21:16 kados the point is 21:16 kados anyway, I digress 21:16 thd kados: Implementing FRBR is very inefficient. Variant data problems in legacy records and too many CPU cycles for processing. 21:16 kados true 21:15 chris thats only about 2 weeks old in library time tho :-) 21:15 kados with not a single app using it? 21:15 kados and FRBR's what, 11 years old? 21:15 thd kados: That seems close. 21:15 kados I thought it was mainly for searching 21:15 kados i didn't think FRBR addressed managing the movement of items 21:14 kados thd: we're not talkinng about FRBR yet :-) 21:14 thd kados: Karen speaks about the possibility of a system designed for supporting FRBR efficiently but she has no design herself. 21:13 chris thats my understanding 21:13 kados right? 21:13 kados keeping track of the movement of items 21:13 kados holdings are used mainly for one thing 21:12 thd kados: Karen Coyle wants to kill MARC but many of her potential allies are intimidated by the LC superpower with the veto. 21:12 kados well, as I see it 21:11 thd kados: Yet, without spending a significant amount of time how will you find a general problem that includes all the issues for MARC and everything else. 21:09 thd kados: Certainly, a fresh approach to any problem with the advantage of knowing the work that has gone before can produce a better solution. 21:08 thd kados: Serials holdings can be very complex. 21:07 kados then have a way to configure the holdings mappings to a specific holdings format 21:07 kados we need to compile some general purpose assumptions about holdings 21:07 kados thd: which is to hardcode it into the scripts, templates and database 21:06 kados thd: but I don't want to make the mistake we made last time with MARC support 21:06 kados thd: so we'll give them those 21:06 thd kados: Most of you customers and the potential ones with the most money to spend will want MARC holdings. 21:06 kados we need to also abstract our record editor's underlying format 21:05 kados so we've made a huge leap forward there 21:05 chris yep 21:05 kados and Koha will find them 21:05 kados ie, I can put in dublin core records into Koha RIGHT NOW 21:05 chris yeah 21:04 kados in the same way we're abstracting our searching format with Zebra 21:04 thd chris: there are over 50 variants but there has been significant unification of the variant formats so that is much less of a problem now. 21:04 chris im not saying it was always broken, just it is now :) 21:04 kados we need a way to abstract our holdings format 21:04 kados so ... more thinking out loud 21:04 chris its still broken 21:03 chris yep 21:03 thd chris: machines know what to do if th programmers tell them how. 21:03 kados true 21:03 chris not 12 different standards that arent standard 21:03 chris when it was a standard 21:02 chris im sure it was a much better thing 21:02 kados chris++ 21:02 thd chris: That is exactly my point about hiding it. 21:02 chris they broke it 21:02 kados machines often don't know what to do with parts of it 21:02 chris but because they do 21:02 kados well ... but also 21:02 chris :) 21:02 chris ever 21:02 chris ever 21:02 chris humans arent supposed to see it 21:02 chris is that everyone forgets the M is Machine 21:02 thd kados: MARC in XML is very extensible. 21:01 chris my main beef with MARC 21:01 thd kados: It was designed when every byte was precious. 21:00 chris ie its purpose has been overloaded and hence its no longer as easy 21:00 kados thd: it's not a very wel-though out storage format 21:00 kados thd: that's only partly true 21:00 chris now people want to use it catalog in 21:00 thd kados: However, MARC has the most carefully thought out solutions to the most complex problems of information records. 20:59 chris MARC was designed for the ease 20:59 chris i think you need to change that too 20:59 kados thd: exactly 20:59 thd kados: Koha should support many different record types. 20:58 kados thd: by a bunch of loony academics 20:58 thd kados: MARC is designed for the ease of record exchange and detailed careful data recording. 20:57 kados but I don't want to be tied to it 20:57 kados in Koha 20:57 kados I want to support it 20:57 kados there's no backing down on that point :-) 20:57 thd kados: Even paul's record editor does much for that. 20:57 kados marc is evil 20:56 thd kados: Hiding the complexity of MARC from the user can readily be done. 20:55 thd kados: I do not mean the ones that want MARC to be hidden. 20:55 thd kados: How many libraries are there in the world that do not want MARC? 20:53 kados because there are instances when libraries or other orgs will not want to use MARC bib records 20:53 kados we don't want Koha's holdings to be _tied_ to MARC standard holdings 20:22 thd kados: Most of the text was still in my head. I could reconstitute it. 20:22 kados be right back 20:21 kados thd: still have it? 20:21 thd kados: I had started a message to you about supporting MARC 21 and now also UNIMARC standard holdings when you were driving off to meet with a law library client.. 20:19 thd kados: They announced at a conference at the beginning of last year that they would be using the separate MARC 21 holdings format records. 20:18 kados and what advantages would it afford us? 20:18 kados yea, so ... how hard could it be really? 20:17 kados I thought they were converting away from it 20:17 kados ahh ... they are converting _to_ MARC ... 20:17 kados it can't really be that complicated 20:17 thd kados: OCLC should now be converting to the MARC holdings format according to their intention announced at the beginning of last year. 20:17 kados what would be involved in getting MARC compliant holdings into Koha? 20:16 kados well ... before I continue down that thread 20:16 kados I can see some advantanges 20:16 kados thd: I'm not 100% convinced we need standard MARC holdings 20:16 kados thd: well ... OCLC doesn't :-) 20:15 thd kados: Of course Koha should head to support MARC standard holdings. Even recommendation 995 support is an adaptation rather than a fully compliant implementation. 20:14 kados yep 20:13 thd kados: What might also be good is to know what software does not use MARC 21 holdings format or uses nonstandard holdings data in current versions. 20:10 thd kados: I will see if paul has an idea tomorrow. 20:09 thd kados: The largest vendors support UNIMARC in some manner so they might have French localised versions. 20:08 thd kados: Unless there is a French branch or subsidiary of a non-French company. 20:07 kados thd: I could be wrong though 20:07 kados thd: I think Dynix does 20:07 thd kados: French systems only. It would not be very meaningful to your customers. 20:05 kados thd: some 'name dropping' would be nice there :-) 20:05 kados thd: in your new version of the holdings doc, it would be interesting to say what other systems base their holdings on recomendation 995 19:39 kados walk even 19:29 kados thd: feel free to bill LibLime :-) 19:28 kados thd: sure :-) 19:28 thd kados: I would be eager to fix the default framework for MARC 21. Would it help me pay my bills? 19:28 kados :-) 19:27 kados thd: feel free to create a new valid framework on LibLime's demo 19:27 kados wonderful 19:27 kados ahh ... has it 19:27 thd kados: Ordering has been mostly fixed. There are still correctable issues relating to presentation. 19:26 kados the subfield ordering thing ... that's tricky 19:26 kados that fixes those issues 19:26 kados ok ... so we need to make sure that 2.2.6 has a better default framework 19:25 thd kados: Only if the user audits the framework and validates everything to add what is missing. 19:25 kados thd: so it's more complicated than just fixing the default framework 19:25 thd kados: exactly 19:25 kados thd: but we _do_ have that now! 19:24 kados thd: ie not preserving it? 19:24 thd kados: The absence of 000-008 in the frameworks makes Koha look very poor to anyone who investigates and causes data loss when editing records. 19:24 kados thd: don't we still have problems with subfield ordering? 19:23 kados hmmm 19:22 kados right 19:22 thd kados: It is still an issue for a record editor that uses the frameworks and also for the several months before 3.X is stable enough for production use.a 19:21 kados thd: at least not for searching 19:21 kados thd: I don't think it will 19:21 kados thd: will that even be an issue now that we've got zebra? 19:20 thd kados: Steven complains about the non-repeatable subfields being set by default to repeatable for 650 etc. 19:19 thd kados: I could build one easily.. It would require a few days to carefully audit everything. 19:18 thd kados: Even the UNIMARC default framework is full of errors. 19:17 thd kados: To avoid the problems that Steven complains about a valid complete default MARC framework for MARC 21 would be needed. 19:16 thd kados: The holdings.html doc has a careless mistake in the mapping but I do have an almost finished revision that is much more comprehensive. 19:14 kados thd: in your opinion, if we implemented your recommendations in the holdings.html doc, would folks like stephen stop complaining? 19:14 thd chris: That is also the most turgid writing that I have ever produced. At least it serves some useful purpose. 19:13 audrey and clearing my fogged brain:) 19:13 audrey thanks, will be looking at all recent information 19:13 chris thats a good explanation thd 19:12 thd audrey: MARC Koha works with the whole call number in one local use subfield. 19:12 audrey why not just one big, changeable field? 19:11 audrey why does marc divide the info on back of books (the sticker bits) into different subfields? 19:11 thd chris: I have an unfinished revision for that document proposing changes. 19:10 chris i thought it was just something olwen and I thought up 19:10 chris cool 19:10 chris ohh is that what its modelled on 19:10 audrey not for status purposes 19:10 audrey classification details are 19:09 audrey item management is not focus at moment 19:09 thd audrey: This might help confuse you more but explains the origin of Koha item management. http://www.kohadocs.org/holdings.html 19:09 kados nope 19:09 audrey not looking for that with classification #'s. 19:08 kados checked out, on reserve, lost, etc. 19:08 kados audrey: managing basically means changing the status of an item 19:08 audrey right 19:08 kados audrey: barcode, item type are two 19:08 kados audrey: well ... there are a few bits of information that Koha pays attention to with items 19:07 kados including Dynix 19:07 audrey what is "manage"? 19:07 kados lots of systems do actually 19:07 thd audrey: you can place the whole MARC record inside Koha but Koha uses a local use field for managing items. 19:07 kados where Koha stores item info currently 19:07 audrey hmm 19:06 kados audrey: though we can nab stuff from the 852 and put it in local use fields in the 900s 19:06 audrey but is mapped and searchable, right? 19:06 thd audrey: Actually, Koha does it that way. 19:06 kados audrey: naw ... not for managing the items :( 19:05 kados thd: lets talk about that ... how nontrivial? 19:05 audrey koha already uses 852. i saw it yesterday 19:05 thd kados: using 852 would require nontrivial changes to how Koha manages items. 19:05 audrey why not in only one subfield? 19:05 audrey am looking at the reasons for prefixes, dewey, cutter--why split between different subfields? 19:04 kados audrey: etc etc. 19:04 kados audrey: and spread across multiple buildings and floors 19:04 kados audrey: academic libraries use it too ... because their collections are also often quite large 19:04 thd audrey: prefix use is common. Suffix use is rare. 19:03 kados audrey: based on their specific needs 19:03 kados audrey: it's really an internal taxonomy for LOC 19:03 kados audrey: exactly 19:03 audrey so the suffix subfield is really because they have departments inside of buildings 19:03 kados thd: (even with zebra?) 19:02 thd audrey: Koha does not use 852 internally. I have a plan but that would require quite a bit of work. 19:02 kados audrey: they have a serious amount of space 19:02 kados audrey: well i have ... it's HUGE! 19:02 audrey in DC, but not at LC 19:02 audrey no 19:02 kados audrey: have you ever been there? 19:02 audrey looking at it now:) will see if it clears the fog:) 19:02 kados audrey: is that it was designed by the library of congress :-) 19:01 kados audrey: the thing to keep in mind with LOC classification 19:01 kados audrey: thd beat me too it with the link 19:00 thd audrey: http://www.itsmarc.com/crs/Bib1424.htm is actually the complete LC documentation for 852 with extra HTML. 19:00 audrey how much difference is there? 18:59 audrey MARC in general, right now 18:59 kados cause there is a difference 18:59 kados are you looking at Koha's MARC or MARC in general 18:59 audrey its subfield says its just for dewey 18:58 audrey the dewey number is easy 18:58 audrey i want more information about why the classification bits go into each subfield 18:58 kados audrey: what then? 18:58 audrey no 18:57 kados audrey: you're confused about classification? 18:57 audrey I've looked at them, and am still confused 18:57 thd audrey: The examples in the LC documentation are the best clue. 18:57 audrey and goes into more than 4 MARC 852 subfields 18:57 audrey the classification number is created from 4/more bits of information 18:56 kados let me pull up a record 18:56 audrey great 18:56 audrey thanks for the recommendations:) 18:56 kados audrey: I can explain them to you 18:56 audrey more detailed than introductory 18:56 audrey they still confuse me 18:55 audrey I need explanation of reason for subfields in 852 field 18:55 thd audrey: This is an excellent brief guide: http://www.loc.gov/marc/umb/ 18:55 kados audrey: http://www.loc.gov/marc/umb/um01to06.html 18:55 kados audrey: this is my favorite intro ... nice and simple: 18:55 audrey have the loc one already 18:54 thd audrey: http://www.loc.gov/marc/ 18:54 thd audrey: http://www.itsmarc.com/crs/CRS0000.htm 18:53 audrey left the card in US 18:53 audrey I don't have a card for KnowItNow at the moment 18:53 kados audrey: sure ... just a sec 18:53 audrey I am still searching for one. 18:52 audrey kados: do you know a website about Marc? with lots of details and examples 18:52 thd kados: full text indexers have no terrible problem. 18:52 kados it's absurd 18:52 kados if $k follows $b it's subtitle if it follows $a it's title 18:51 kados can deal with MARCs notion of subfield order 18:51 thd kados: I looked for a supposed reply from paul but found none. 18:51 kados no indexer on the planet 18:51 kados for instance 18:51 kados well ... the problem with MARC as I see it is that it doesn't live up to its name as a machine readable record format 18:50 thd kados: Yet, after 4 decades that is a little problematic. 18:49 thd kados: The problem with MARC is that in order to be a very meaningful standard LC has been very resistant to changes for mere data format fashion. 18:48 kados but believe me, koha developers do care about it 18:47 kados MARC is a terrible standard 18:47 kados of course 18:47 thd kados: He cares greatly about library standards as do I and every one working on Koha should. 18:47 kados thd: what is he referring to? 18:47 kados thd: any comments on his statement about the 650? 18:46 kados right 18:46 thd kados: I have read. I have corresponded with Steven in the past his details about Koha are not current enough sometimes. 18:36 thd kados: still reading 18:35 kados thd: i'd like to talk about how best to respond 18:35 kados thd: after you read it ping me 18:32 kados did you see stephen balkits latest email? 18:31 thd yes kados 18:31 kados thd: you still around? 17:54 jungle : ) 17:54 jungle hi to all hope everything is good 17:54 jungle hello 17:44 thd dce: That is probably a good price to even save you the trouble of sending a query to a Z39.50 server. What does Follett charge for the books themselves in relation to the publisher's suggested retail price? 17:39 dce thd: The MARC records from Follett are $0.26/record. 17:19 thd kados: A change is eventually needed in how the frameworks function so that plugins can operate on discrete parts of fixed fields that correspond to an individual value. Fixed fields would then be much easier to manage. 17:14 thd kados: It is much less work for the cataloguer to do things once for the whole record but much more work for the programmer. Yet, the parallel reverse script for reading is needed for even copy catalogued records. 17:09 thd kados: A sophisticated plugin that added values to multiple fields at once would be nice though, that could be the reverse of a complex general and special material type reading script. 17:06 chris it might still get sunny .. its threatening too 17:06 rosa here too 17:06 thd kados: The real problems are reading the meaning of the coded and other values. Writing them to the individual fields is is easy for a plugin. 17:06 chris it was yesterday 17:06 rosa bummer. I thought it was going to be beautiful 17:04 chris sorry rosa, middling here too 17:04 thd good morning chris and rosa 16:55 rosa It's very middling here 16:55 rosa Welly, even 16:55 rosa how's the weather in elly? 16:54 rosa all 16:54 rosa morning 16:53 chris morning thd and kados 16:53 chris morning rosa 16:48 thd kados: The problem shown by that table is that you need more than one fixed field for even distinguishing whether language material is a serial or a book. 16:46 kados thd: thanks 16:46 thd kados: If you want a simple plugin for the material types that table and the related fixed field documentation will help you write a plugin. 16:44 thd kados: Proper material typing is too inefficient to be done at query time. 16:43 thd kados: I had started writing some Perl code months ago for determining material types from the MARC record but I aborted that effort with the number of fields involved in either general or special material designation. The question is complex and needs to be stored in a separate local use MARC field for Koha 3.0. 16:41 kados with a plugin that auto-filled the proper fixed field 16:41 kados I could easily create a framework for each material 16:41 kados yea, that is useful 16:40 thd kados: This table is useful: http://www.itsmarc.com/crs/Bib0020.htm 16:34 thd kados: look at the MOD mapping for material types also the complete MARC 21 Bibliographic manual hosted on TLC. 16:34 kados yea, suppose a plugin would work 16:32 thd kados: Look at paul's plugin for UNIMARC 100. 16:31 thd kados: You could write a plugin for those to use in the Koha record editor, 16:30 kados material type and the like 16:29 kados I just had someone ask about koha's ability to handle fixed-field itemtypes 16:28 thd kados: what does your question mean? Paul does have a plugin for modifying the most used UNIMARC fixed field. UNIMARC 100 $a is a fixed subfield that along with other 1XX works similarly to MARC 21 005-008 fixed fields. 15:55 thd kados: I cannot see how anyone drinks coffee ever but I thought it tasted foul when I sipped it once and forever found it easy to say no. 15:53 thd kados: How differently do you want to store the fixed fields? 15:47 kados thd: I haven't played with the frameworks much 15:47 kados thd: do you happen to know if it's possible to create frameworks within Koha that store fixed fields differently? 15:46 kados I don't drink coffee every day 15:46 kados heh 15:46 thd kados: just say no 15:45 kados too much coffee today :-) 15:44 kados that even 15:44 kados cancelled taht is :-) 15:44 thd dce: my vision is focused now and I see that you are dce not dice as I had used. 15:40 thd dice: I am interested in knowing how that value is priced or if it is unavoidably built into the service. How does the service price differ from the publisher's suggested retail price. 15:37 thd s/you/your/ 15:36 thd thd: There may still be the advantage of convenience, known quality standard, and actually having a record that is hard to find elsewhere by obtaining records from you vendor. 15:36 dce I'll lurk here for a while waiting for the librarian to give me more details then I'll let you know. 15:35 dce thanks. 15:34 dce thad: cool 15:34 thd dice: Any other library with a Z39.50 server, which maybe any other library from a neighbour to the national library in another country. 15:32 dce thad: copy catalogue from where over Z39.50? I'm afraid that I have very little experience with the library side of things but a lot on the tech side. 15:32 thd dice: I am interested in knowing what the price differential is between purchasing books without records supplied and purchasing with records supplied. 15:30 thd dice: MARC records may always be imported at any time but you are not dependent upon Follett records. You could also copy catalogue over Z39.50. 15:29 dce apparently they can provide the records in more than one format. I'll get more details 15:28 kados dce: to map your holdings data to Koha's internal structure for holdings) 15:28 kados dce: (though it may require some customization of the import file) 15:27 dce thd, I can ask 15:27 kados dce: (In that case, you can import MARC directly into Koha from the disk ... ) 15:27 dce We are looking at using Koha in a new school opening this fall and they want to know if they will still be able to import that info. 15:27 thd dice: Do those records cost extra? 15:27 dce kados, when ordering books from Follett - http://www.flr.follett.com/ - our library staff has the option of getting a disk with MARC records. 15:26 kados dce: if you'd like to talk more about it later, let me know 15:26 kados dce: but if we had a sponsor for such a feature it could be done very quickly 15:26 kados dce: if you mean integrating the order process within Koha's acquisition module, noone has done that to date 15:26 thd kados: I await your ping 15:25 kados thd: I'll ping you when I am available to talk 15:25 kados thd: probably around 3:00 or so 15:25 kados dce: could you expand on what you mean by that? 15:25 thd kados: after your call would be when? 15:24 kados thd: yes, but I have a conference call soon I must prepare for 15:24 thd kados: I am awake again, are you? 15:19 dce has anyone here dealt with importing MARC data from Follett when ordering books from them? (Not migrating from Follett software.) 15:00 dweezil19 thanks, I've got to run now 14:57 dweezil19 it is an ldap server so that should work (i think!) 14:57 dweezil19 i guess so 14:57 kados dweezil19: can you use that with active directory? 14:56 kados dweezil19: Koha has support for LDAP 14:56 kados dweezil19: hypothetically, yes 14:20 dweezil19 Is it possible to authenticate via active directory? 14:04 kados dweezil19: sure 13:58 dweezil19 can i ask a quick question about koha? 13:56 kados hey thd 13:52 kados owen: not sure if support for it is built in to the scripts yet 13:52 kados owen: the syspref's already in rel_2_2 13:52 kados owen: yep 13:34 owen kados: is the plan to create an opac system preference for color stylesheet like the intranet has? 12:39 jungle hello people