Time  Nick     Message
12:00 hdl      2. you input the biblio for the General serial
12:00 hdl      For instance : Official bulletin on the internet
12:01 hdl      Then you go to serials in the catalogue to open a new subscription.
12:02 hdl      Link biblio with supplier provide date of begining, period or number of serials to be received with the subscription.
12:02 hdl      Make the Numbering Formula.
12:04 hdl      And there you are... You can know wich serial you are waiting, which have been recieved.... IN OPAC as well as in Intranet if FullSubscriptionHistory is set.
12:04 kados    interesting
12:04 hdl      Doss that answer your demand ?
12:04 kados    I think so
12:04 kados    so when I search on opac I'll see what issues are avaialble
12:04 hdl      Yes.
12:05 kados    but what if I've got a bunch of serials with each issue in an 852 tag (for holdings)
12:05 thd      hdl: what about individual issue barcoding?
12:05 kados    hdl: is there a way to map those issues (in 852) to the serials in Koha?
12:05 kados    hdl: (or they could be in 952)
12:06 hdl      They are already stored in serial table.
12:06 kados    hdl: ? if I import a new MARC record it's already in the serial table?
12:07 kados    hdl: I'm talking about a library migrating to Koha using MARC files that have one record for each serial with many many 852 fields for the issues
12:08 kados    is there a way to map the issues to the serials table in Koha?
12:09 kados    does that make sense? ;-)
12:09 hdl      kados : No. I said that serial table was used for subscription management....!!!!
12:09 hdl      kados : But I did NOT say it created a MARC record.
12:09 hdl      kados : but what if I've got a bunch of serials with each issue in an 852 tag (for holdings) I didnot understand.
12:09 hdl      kados : is there a way to map the issues to the serials table in Koha? I think there should be but need some work.
12:09 kados    ahh
12:10 thd      hdl: If I know the library to which kados is referring, then the 852 fields for each issue have the volume and issue number in each 852 field.
12:10 kados    in your opinion is it good to store issues in 852 fields ? or should each issue have it's own marc record
12:10 kados    thd: right
12:10 hdl      I think that it all depends on the library.
12:11 kados    thd: what do you think?
12:11 thd      kados: Why would each issue have its own MARC record?
12:11 kados    thd: isn't it risky to store issues in 852 fields since there's an upper limmit to how large a MARC record can be?
12:11 hdl      If serial searching should lists numbers, you should make a new MARC record for each number.
12:12 kados    thd: how is this normally handled?
12:12 hdl      If it is more a question to know if there is a subscription to this serial... Then you should map to 852 field.
12:12 kados    thd: (within marc)
12:12 thd      kados: the problem is I think how the issues are stored,
12:13 thd      kados:  Look at the poorly written yet useful migration document that I wrote.
12:14 thd      kados: It is on kohadocs.org now in section 2.2 for migration.
12:15 kados    thd: yep ... reading the serials holdings section now
12:17 kados    thd: how do you suggest addressing the issue? ;-)
12:17 thd      kados: The problem with the library to which you are referring is that they have put each issue in a separate 852 field rather than represent the combined information in fewer repeating fields.
12:18 kados    thd: right ... so they're doing it wrong then ;-)
12:19 thd      kados: Well they are using the limitations of their existing software perhaps.
12:19 kados    thd: I still think you'd run into a size limitation with the MARC record
12:19 kados    thd: for some serials
12:20 kados    thd: I'm wondering whether most libraries use MARC for serials or whether they do them using some kind of internal serials module like Koha's
12:20 thd      kados: In a fully MARC compliant database, multiple records can be linked to one another to avoid to much information in one record.
12:20 hdl      kados thd : In UNIMARC you can have a link down i.e. from serial to issues OR/AND a link UP from issues onto serial father.
12:21 hdl      Ther should be the same in US-MARC.
12:21 thd      hdl: MARC 21 has the same provision for linking records to parent , etc. records.
12:22 kados    does Koha support this?
12:22 hdl      Yes...
12:22 kados    I wasn't aware if it ... how does it work?
12:22 hdl      I did unimarc_field_4XX plugin.
12:22 hdl      Well....
12:23 hdl      It just get the biblionumber of the father or son into the required field.
12:24 thd      kados: I had suggested this to Owen a month ago for what he was interested in doing for simply creating a temporary serial record for an issue that would be discarded after a brief period.
12:24 hdl      But then, searching through or getting FULL inheritance. That is not implemented.
12:25 thd      kados, hdl : There is coding to be done for this problem, as Owen replied to me.
12:25 hdl      which one ?
12:26 thd      hdl: I mean implementing record linking more completely in Koha.
12:29 hdl      Yes. But would need a detailed spec to stick closer to it and to know what is really needed.
12:29 thd      On many levels there are problems.  In the OPAC, if every issue has its own record and is not discarded the borrower does not want to browse through hundreds of issues to find the right one.
12:35 owen     http://koha.org/cgi-bin/logs.pl?user=&channel=&action=&text=&user-ddl=&channel-ddl=&action-ddl=&startdate=14%2F07%2F05&enddate=15%2F07%2F05&saveas=&save=save&saved_query=Town%2520Hall%2520Meeting+2020+20May+2020
12:39 owen     Koha's serials management is pretty much disconnected from circulation.
12:39 kados    from MARC too unless I'm mistaken
12:41 hdl      No Kados, youre not, unfortunately.
12:43 thd      kados: It is connected to MARC if you create a barcode for each item and you have your barcodes linked to a MARC subfield is it not?
12:43 owen     Serials management doesn't have any way of connecting individual issues to individual items
12:44 kados    hdl: have a good evening
12:44 hdl      thx.
12:44 kados    hdl: we can talk about this again ;-)
12:45 hdl      When you want. It is interesting.
12:46 kados    owen: what if there was a 'serials framework'
12:46 thd      good night hdl
12:46 kados    owen: does that make sense?
12:46 owen     Heh...that's what prompted my discussion with thd about serials.
12:46 kados    ahh :-)
12:47 owen     I was telling him how libraries might like to have a framework especially for serials.
12:47 thd      I am about to paste it.
12:47 kados    it would also be nice if you could 'flag' a record and have it use a specific framework
12:47 kados    when you're importing
12:47 thd      relatively short.
12:47 thd      16/07/05 03:49:11+-5	<thd:#koha>	owen: the ideal way to address the disposable serials issue is with a 772 linking field
12:47 thd      16/07/05 03:50:19+-5	<thd:#koha>	I imagine there is some functionality one could borrow from the UNIMARC side
12:47 thd      16/07/05 03:51:19+-5	<owen:#koha>	Can you elaborate?
12:48 thd      16/07/05 03:52:13+-5	<thd:#koha>	A well designed 77X plugin should copy required information from a linked record and then you would only need to modify a few subfields in 3 or 4 fields
12:48 thd      16/07/05 03:54:17+-5	<thd:#koha>	So you would get all the bibliographic information that was the same for every issue of your serial copied from the other record and you would add the issue and holdings data.
12:48 thd      16/07/05 03:55:29+-5	<thd:#koha>	This would be a MARC21 compliant solution :)
12:48 thd      16/07/05 04:01:38+-5	<thd:#koha>	The linking fields are 76X - 79X. I am trying to fetch an example for you but my system is thrashing a bit.
12:48 thd      16/07/05 04:02:10+-5	* indradg:#koha joined
12:48 thd      16/07/05 04:03:43+-5	<thd:#koha>	245 04$aThe Post boy.$nNumb. 2436, from Thursday December 21 to Saturday December 23, 1710.
12:48 thd      16/07/05 04:03:43+-5	<thd:#koha>	772 1#$7unas$aPost boy (London, England)$w(OCoLC)1234567
12:48 thd      16/07/05 04:04:05+-5	<thd:#koha>	http://www.itsmarc.com/crs/Bib1390.htm
12:48 thd      16/07/05 04:06:04+-5	<thd:#koha>	owen: In the example above, 772 $w is the control number for the linked record on your system and 245 $n has been added for the issue.
12:48 thd      16/07/05 04:08:12+-5	<owen:#koha>	Sounds great in theory :)
12:48 thd      16/07/05 04:08:23+-5	<thd:#koha>	owen: 245 $a , 772 $a, and much else could have been copied by a plugin from the linked record
12:48 thd      16/07/05 04:12:31+-5	<thd:#koha>	owen: I know there may be a more efficient way to solve your immediate problem with less effort but this standards compliant method would be reusable even for permanent records and other institutions where people go crazy over MARC standards :)
12:48 thd      16/07/05 04:23:48+-5	<owen:#koha>	Something to be taken up with the developers...
12:50 owen     I don't see why the serials management process could integrate an item-creation routine each time an issue was received.  At that point you could enter barcode information, etc.
12:50 owen     could->couldn't
12:51 thd      owen: exactly
12:52 owen     kados: what other kinds of things is this library looking for in serials?
12:53 thd      I assume that the procedure that hdl was referring to does not allow an individual barcode per issue.
12:53 kados    owen: I really have no idea
12:53 kados    owen: it's the law library that contacted me recently
12:53 kados    owen: that got me thinking about serials
12:53 kados    owen: because they said they'd have to look at the serials support closely
12:53 kados    owen: I said I'd look into it and get back to them
12:54 kados    owen: what I'll probably do is get a list of things they want to do
12:54 kados    owen: and report back to koha-devel
12:54 thd      kados: So you were not referring to the church library?
12:54 owen     You and I probably have similarly-lacking knowledge of serials management, since it's not something we do at NPL
12:54 kados    thd: no
12:54 kados    owen: exactly
12:54 kados    owen: all it takes is one client though ;-)
12:55 kados    owen: functional abstraction rocks ;-)
12:55 kados    owen: except when it doesnt ;-)
12:55 thd      kados: what is their existing software?
12:55 kados    thd: not sure yet
12:56 kados    thd: but it's one of the cheaper ILS solutions ... not one of the big guys
12:56 kados    thd: probably follett or Sagebrush
12:57 kados    nope ... dont' have it recorded
12:59 thd      kados: is the client near enough to you so that you can see their existing system in action?
13:00 kados    thd: yea ...
13:00 kados    thd: same state even ;-)
13:01 thd      kados: well, it would be good to have some real information to see how it works with what they are already using.
13:01 kados    thd: yep
13:01 kados    thd: I'm planning to do just that
13:03 thd      kados: about your record matching problem over Z39.50
13:03 kados    thd: yes?
13:04 thd      kados: you can see the same problem in a smaller scale in Koha.
13:04 thd      kados: NPL has few enough records that it should not appear much.
13:06 thd      kados: In Koha all the various fiends that might contain author information can be linked together for an author search.
13:07 thd      kados: in MARC 21 that is 100, 245 $c 700, etc.
13:07 thd      kados: there are many more I just simplify.
13:08 kados    right
13:08 thd      kados: so if you search for John Smith,
13:09 thd      you may find a record that has John Knox and James Smith.
13:09 kados    right
13:10 thd      kados: The larger the database is the more likely such mismatches are to be a problem.
13:12 thd      kados: This goes to much of what Karen was trying to tell me about how the number and size of indexes increase when you do things right.
13:14 thd      kados: The match I gave above should never happen because the query was being run against too many indexes combined as if they were one.
13:15 thd      kados: combining the indexes in that way can improve search performance only at the price of lost accuracy.
13:17 thd      kados: Ideally, in a conventional system, you want to search 100 separately from 700 and only report a match if any one of them has a match.
13:17 thd      kados: not if there is a match collectively on all the terms indexed.
13:19 thd      kados: 245 $c , if included is a problem because the statement of responsibility there puts all the authors in one subfield.
13:20 thd      kados: sorry, temporarily distracted by a phone call one moment.
13:23 thd      kados: UNIMARC provides for subfield separation for last names in the author fields.
13:25 thd      kados: forcing the user to specify author queries exactly to match one combined index can be problematic.
13:27 thd      kados: so for the OPAC, conventional systems check each relevant index separately consuming many CPU cycles.
13:27 thd      kados: The exact problem you were having with Z39.50 is a little different.
13:30 thd      kados: Whether the system saves some cycles on Z39.50 searches at LC and uses a combined index like Koha I have not tested for yet.
13:32 thd      kados: however, I have found both author and title matches to personal name subjects, 600, and formatted tables of contents, 505, for authors.
13:33 thd      kados: titles words also match formatted tables of contents, 505, for LC Z39.50.
13:34 thd      kados: to get the most relevant matches over z39.50 you have to adjust the query and the search attribute syntax.
13:35 kados    thd: what do you suggest?
13:36 thd      kados: most importantly for your purpose n identifying exactly matching records tiles must be a phrase search.
13:37 thd      Yaz supports two useful types of search attribute in addition to all the others.
13:38 thd      I have not looked to see how it is specified in Net::Z39.50
13:42 thd      kados: One useful way of supplying a phrase search in Yaz is s=pw .
13:43 thd      kados: This gives you a phrase search for multiple tokens in the query.
13:45 thd      kados: I do not remember the numeric equivalent but I seemed to have trouble with numeric equivalents sometimes.
13:46 thd      kados: you need to search on multiple elements as well.
13:48 thd      kados: author-title as a single search covers too many fields and cannot work as a phrase search with both author and title if you use both, unless my information misses something about author-title searches.
13:51 thd      kados: you need u=4 or 1=4 for title so for your title phrase from 245$a omitting the leading article you would have u=4 s=pw as attributes.
13:54 thd      kados: In the same query you should have the author's last name form 100 $a or 700 $a etc. if 100 etc. is missing to use with u =1003 attributes.
13:57 thd      kados: that might need some refinement if corporate authorship is an issue.  I do not remember at this moment if u=1003 is all authors or only personal names.
13:58 thd      kados: Then if there are multiple manifestations of the same work or if that is not accurate enough otherwise you need to go next to publisher.
14:00 thd      kados: Before I forget to mention it, the reason I advise against using first name for authors is you do not know whether you have authority controlled names.
14:02 kados    thd: makes sense
14:02 thd      kados: you might try experimenting with last name and first initial as a string if that could be made to work with LC over Z39.50.  that would be a string search with a numbered search attribute not a phrase search.
14:03 thd      kados: So publishers are very tricky because there is no authority control for publisher names.
14:05 thd      kados: you want to identify the Wiley in John Wiley and Sons and search only for that if you can.
14:08 thd      kados: certainly for publisher you can eliminate common words like 'and', 'publishers', 'university', 'press', etc.
14:16 thd      kados: so with 260 $b extract the most important words to submit with u=1018 s=al attributes. s=al is for all words.
14:17 thd      sorry, that my response time is slowing my system has been thrashing for a while today due to a memory leak in Flash ads in Firefox.
14:18 kados    :-)
14:20 thd      kados: Dates or edition information might be used but how they are recorded in records is too variable to be relied upon over Z39.50.
14:21 thd      260 $c may have only a reprint date in your record or the Z39.50 target record etc.
14:23 thd      kados: For the past two weeks I have been intensely researching how to identify the true duplicates that may be left in a result set of potential candidates.
14:24 thd      kados: I have seen the dirtiest microfilm for an academic journal that I have ever encountered.
14:24 thd      kados: The good detailed published research seems to be quite old.
14:25 thd      kados: All my interlibrary loan requests are still outstanding.
14:27 thd      kados: Identifying the true duplicates in a candidate result set gets very tricky quickly because of the lack of absolute uniformity of cataloguing practice between cataloguers and over time and following AACR2 etc.
14:29 thd      kados: In the UNIMARC world they at least have the luxury of authority control over publisher names.
14:34 thd      kados: If you can identify duplicates reliably over z39.50 and after you could enhance records automatically by merging the fields in records obtained targets that had more complete records or merely additional encoded fields.
14:36 thd      kados: MELVYL used to do this and it is the major focus of section 3 of my yet to be sent email to you.
14:37 kados    thd: you ever going to send it ? ;-)
14:38 thd      kados: two weeks ago I had it very close to finished but I could not stay awake long enough to proof read it.
14:38 thd      kados: I mean I could not see straight to proof read it after I had finished writing :)
14:39 thd      kados: I want to add the appropriate references to the published research as well.
14:41 thd      kados: Every catalogue should have the most complete records possible but even OCLC master records do not achieve this for many cases.
14:43 kados    thd: that's because they're written by librarians ;-)
14:44 kados    thd: s/written/contributed/
14:44 thd      kados: I have not addressed the old MELVYL merged record issue with Karen we mostly talked about her ideas about making systems more efficient so that you could actually start to do FRBR efficiently instead of spending too many CPU cycles for a less than effective result.
14:45 thd      kados: OCLC pays libraries for both contributing original and enhancing existing records.
14:45 kados    thd: but all records come from librarians
14:46 thd      kados: The payments are obviously not high enough.
14:47 thd      kados: you have added routines to add information to the OPAC contained in records that does not come from libraries.
14:48 kados    thd: right
14:48 thd      kados: Amazon gets information from distributors and publishers, as well as creating their own.
14:48 kados    thd: does that mean I should get paid the same as a cataloger who manually updated those records? ;-)
14:49 kados    thd: pipe dreaming ;-)
14:49 thd      more kados, more;
14:49 kados    thd: hehe
14:49 thd      ;-)
14:49 kados    thd: yea ... because I can do it in an hour instead of a year ;-)
14:50 thd      kados: there is a difference though with what you did and what a cataloguer does in enhancing records.
14:53 thd      kados: You need to comply with Amazon's data licensing terms against permanent long term storage of their data.  But a cataloguer enters the enhanced information into a record so that it is searhable.  What you have done is not searchable yet.
14:56 osmoze   thd, have you receved my mail ?
14:57 thd      kados: Information from Amazon should go temporararily into 5XX fields, or whatever fields are appropriate so that it can be searched.
14:59 thd      osmoze: yes thank you I was too tired last night to examine your records but what is your relation to Ecole Normale Supérieure de Cachan ?
14:59 osmoze   nothing ^^
15:00 osmoze   i ve just searching any z3950 in france
15:00 osmoze   for me
15:00 osmoze   i m a little public library, not school library
15:00 thd      osmoze: their records do have the ISO 5426 character set.
15:01 osmoze   yes
15:01 osmoze   that what you are searching for no ?
15:01 thd      osmoze: In their records accented characters are multibyte.
15:02 thd      osmoze: yes that is exactly what I was expecting to find.
15:05 thd      osmoze: Their records have standard UNIMARC character set encoding using ISO 5426 that Koha is not designed to use yet.
15:06 osmoze   hum...and me ?
15:07 osmoze   i don't understand why you say that koha is not designed to use
15:07 osmoze   there is only for characters
15:08 osmoze   so, the fonction of koha not differ
15:08 thd      osmoze: The confusing thing is that BNF would seem to be sending records over Z39.50 in single byte 8859-15 while the encoding given in 100 specifies double byte ISO 5426.
15:09 osmoze   hum...oki :) i understand now :)
15:13 thd      osmove: Koha would need some special routines to change the user query in the OPAC from single byte 8859 to multibyte ISO 5426 for searches on records storing data in ISO 5426.
15:14 thd      osmoze: The opposite would need to be done for displaying the record.
15:16 thd      osmoze: The record editing module would need to have a special ISO 5426 font that may not be possible without extending the functionality of the web browser.
15:17 thd      kados: do you know anything abut adding support for non-standard fonts to a web browser?
15:17 osmoze   hum...oki, but the difference between iso 8859-1(or 15) and iso 5426 ?
15:18 owen     There is some available technology for using 'downloadable' fonts in a web page, but it is limited and not cross-browser friendly
15:19 owen     Why would you need a non-standard font?
15:20 thd      osmoze: ISO 5426 just represents the accents.
15:20 osmoze   yes but, in 8859-15 accents are present
15:23 thd      own: Koha cannot support existing MARC standards without supporting MARC-8 for MARC 21 and ISO 5426 for UNIMARC.  There may be some other character set issues as well but those are the main sets used for Latin characters.
15:24 thd      own: NPL has been cheating which is fine for NPL but that will not do at most libraries with a more diverse collection that is more likely to have records with accents.
15:24 owen     It's not a question of the font in the browser, it's a question of the data stored in the database and the data coming in from the user's keyboard.
15:25 owen     We can help by giving the proper encoding to the web pages and to the database structure
15:25 owen     Does UTF-8 not cover the proper character set?
15:26 thd      owen: A problem is that there is no standard web page encoding for these library character sets.
15:26 owen     What in particular makes it a 'library character set'?
15:27 thd      owen: Well actually converting to 8859 or UTF-8 for display would be necessary.
15:28 thd      owen: and then converting back again.
15:28 owen     What makes this character set necessary for libraries?
15:29 thd      owen: for storing an edited records information in a standards compliant way.
15:29 thd      owen, osmoze: these library character sets were designed before unicode.
15:29 owen     Yes, but what makes the character set suitable for libraries?
15:30 owen     What characters are included that aren't available in other character sets?
15:31 thd      owen: These character sets are difficult to work with and bad in the present time compared to the others.
15:32 thd      owen: Unfortunately, these are the ones supported by library standards, and contained in properly encoded library records.
15:33 owen     Saying the character set is different because it's the one required by library standards doesn't really provide any useful information about the character sets themselves.
15:35 thd      owen: you can use other encodings, but that requires converting the record and changing the coded character set encoding designation in MARC 21 as required for MARC 21 and UNIMARC 100 $a/26-29 etc.
15:36 owen     thd: what makes the library-specific character encoding different from, say, UTF-8?
15:36 thd      owen: sorry, I cannot type fast enough and my system is thrashing.
15:39 thd      owen, osmoze: UTF-8 is well supported outside library records and would be easy to use but it has a different encoding from library records currently.
15:40 genji    hey kados, you still in?
15:42 thd      owen, osmoze: If you encode your records in a superior such as with full unicode for example you will not be able to support exchange of your records with other libraries because their systems are mostly not compliant with full unicode and full unicode is not included in library standards.
15:43 owen     Is there any reason why Koha couldn't just have an export option that allowed the proper encoding?  Is there any reason to actually *store* using this required encoding?  What advantage does the encoding offer?
15:43 osmoze   oki oki ^^
15:44 osmoze   but, export record is often in same encoding no ? i m thinking that my record don't interresting you
15:44 osmoze   no ?
15:44 thd      owen: that would be the best approach.  The exta advantage is only for record exchange and not having to convert for exchange purposes.
15:45 owen     Seems like it's hardly an advantage to 'not have to convert for exchange' if you have to figure out how to re-write Koha to handle an obscure character encoding.
15:46 thd      osmoze: Well I am expecting that your records are interesting simply from the character set encoding being ISO 8859-15 while the 100 $a/28-29  indicates that the encoding should be ISO 5426.
15:48 thd      owen: you are perfectly right about which would be more work and more efficient.
15:49 thd      I mean and which other would be more efficient.
15:52 thd      owen: so koha should move to UTF-8 or full unicode and then convert for import and export.  Either way is still a lot of work to do once.
15:53 owen     I would be surprised if there were many ILSes out there that couldn't import records in non-standard encodings
15:54 thd      owen: You cannot expect that the OPAC user will have UTF-8 or unicode outside the library so conversion would be needed for display and queries but at least standard tools are available for that purpose.
15:55 owen     So you're saying that browser support for UTF-8 shouldn't be expected?
15:57 thd      owen: I think for the browser to support UTF-8 the user has to install UTF-8 font support.  I am not sure that the millions of Windows 98 and less users have the right font support available by default.
15:58 owen     I'm guessing that the worst that will happen is that they'll get some blank characters where a special character should be
15:59 owen     http://www.cafepress.com/nucleartacos.26746952
15:59 thd      owen: The newer the OS the more likely the font support is there without special user effort.  The older the OS the more likely a special effort is needed to want CJK support for example.
16:00 thd      owen: blanks, question marks or strange boxes.
16:00 owen     CJK support?
16:07 thd      owen: Chinese, Japanese, and Korean.  If you have Windows 95 or Windows 98, and you install CJK support or some other support that requires at least UTF-8 then the system supports UTF-8.  Otherwise the system does not support UTF-8 at all.  UTF-8 support could of course be added for only Latin characters but even then requires a special update for most European language users.
16:08 thd      owen: European languages do not need UTF-8 or unicode but CJK requires at least UTF-8.
16:15 thd      owen: everything would work much better with full unicode as the standard for library records but the ALA rejected that for lack of readiness and fear of change a few months ago.
16:19 thd      owen: I am given to understand that major ILS vendors want this but the librarians do not want to carry the one time expense of record conversion.
16:20 thd      s/librarians/library administrators
16:21 thd      owen: The cataloguers want this because it would make their task a little easier.
16:22 osmoze   bye
16:23 thd      good night osmozZz
16:23 genji    aww
16:23 osmozZz  thx :)
16:24 genji    wanted to hash something out with someone.
16:24 genji    owen, you available?
16:24 owen     Yes
16:25 thd-away genji will you be up in 6 hours?
16:25 genji    yup
16:25 genji    will owen?
16:26 thd-away owen is up now though, but I am gone
16:27 genji    right., ill talk to owen.. if i need further help, ill ask in 6 hours.
16:27 genji    anyway, owen... you know how each borrower has flags right?
16:27 owen     Yes
16:28 genji    Right. Now, Kados idea was to expand that idea, and give borrower categories permissions, and have them inherit to the users in their group.
16:30 genji    What you think of that?
16:31 owen     I'm not sure what 'inherit to the users in their group' would mean
16:34 owen     You just mean that each person in a particular category would have the permissions that had been set for that category?
16:34 genji    okay..... If i create a borrower in group "Admin" the admin group will give permissions to the borrower of its category.
16:35 owen     I think that makes a lot of sense.
16:35 owen     It would also set things up so that libraries could think about setting other staff-specific parameters like loan length, etc.
16:37 genji    then, how do you represent permission overrides, in the modify user flags page? in windows, a unset inheritance would be represented by a grey check in a checkbox. But, in web browsers, we can't do that.
16:38 owen     Ah, so you're thinking it should be possible to override the permissions set by a borrower category?
16:41 genji    if the person inside modify flags desires it, yes. but how to represent inheritance?
16:42 owen     Possibly a second column indicating which flags to override? (Of course that assumes you only override something to *on*)
16:43 genji    hmm.
16:50 genji    What if you want to override something to off?
16:52 genji    hmm. this is difficult.
16:54 genji    how does one represent the fact that a flag was overridden, on previous edits?
16:56 owen     What about two columns, one showing group standard flags and one current user flags?
16:57 genji    hmm.. and the group standard flags are solid state, they're just graphical representation?
16:58 owen     Yeah
16:58 genji    so the only time the flags are inherited, is in new borrower mode?
16:59 owen     I'm assuming the flags would be set according to which borrower category you chose when you added the member.  And maybe you couldn't even create a borrower with higher privileges than yourself
17:00 genji    k. laters then.
17:00 owen     I'll check the logs tomorrow
03:12 hdl      kados, thd, owen: what you described at 19:50(French time) yesterday and 4:23 (N-Ztime) on the 16/07 is actually done with unimarc_field_4XX plugin.
03:12 hdl      kados, thd, owen :Thus, One can connect 7XX plugin on the basis of this plugin... (Maybe some changes should be done on the subfield labels)
03:12 hdl      Salut Sylvain.
03:12 Sylvain  salut hdl
03:30 osmoze   bonjour
03:59 hdl      salut osmoze
04:19 hdl      thd : I think that owen is true saying that Koha should now head toward UTF-8 support and then use conversions from and to weird charset only on import/export.
04:19 hdl      That is required with CJK but also with arabic languages... and seemms not so away from Koha.
05:11 hdl      hi indradg
09:43 thd      hdl hello
09:43 hdl      thd hello !
09:43 hdl      nice nap ?
09:43 thd      yes
09:45 thd      hdl: part of what I had said about character set issues yesterday was mistaken.  I need to correct that.
09:46 hdl      On which theme : serials, Z3950 ?
09:46 hdl      both ?
09:48 thd      hdl: I had misinterpreted some documents about the MARC-8 set for MARC 21 where it referred to that set as default and made an improper assumption.
09:49 thd      hdl: MARC 21 has allowed UTF-8 since 2000 but not full unicode since 2000 provided leader, 000, is set to a.
09:50 thd      hdl the existing records in the world are in MARC-8 and require conversion.
09:52 thd      hdl: the reason that full unicode is not allowed yet has been given as ILS system unreadiness.  But the vendors, I have been told are waiting on the standard to require it before implementing full unicode.
09:52 hdl      hi owen.
09:53 owen     Hi
09:53 hdl      Why note having a full Unicode Koha before ?
09:54 hdl      It would be More than great on a marketing point of view when ILS vendors comes to UTF-8.
09:54 thd      hdl: Koha should have full unicode because it is in the UNIMARC standard.
09:55 thd      hdl: If 100 $a/25-27 is set to 50 then the record is full unicode.
09:56 thd      hdl: Of which, UTF-8 is a subset.
09:57 hdl      So : taht leads to choose UTF-8 with transcoding on US-MARC import and export or even only in import ?
09:58 thd      hdl, owen: MARC::Charset provides conversion from MARC-8, thank you again Ed Summers.
09:59 hdl      AHA :D
09:59 hdl      smileglasses
10:00 thd      hdl, owen: There is also a closely related module for converting back into MARC-8.
10:05 thd      hdl, owen: The existing base of UNIMARC records in ISO 5426 can be converted with Encode::MAB2.
10:05 hdl      MAB2 : what is this format ?
10:06 thd      hello osmoze and owen I hope you have read my correction from yesterday above.
10:07 thd      hdl: MAB2 is the German MARC and is ISO 2709 compliant unlike MAB, which had not been.
10:10 thd      hdl, owen, osmoze: changing character sets in the record is supposed to require changing the character set encoding designations in the record.
10:11 thd      hdl, owen, osmoze: Otherwise another system cannot be expected to know how to read exchanged records.
10:12 hdl      thd : this can be managed when exchanging datas.
10:14 thd      hdl: BNF seems to be doing something that may be useful but is certainly not standards compliant in providing UNIMARC records in the non-UNIMARC standard ISO 8859-15 but leaving the record encoding set to ISO 5426.
10:15 hdl      But most of the time, ppl inside the library we serve, either DON'T inform these fields would it be laziness or ignorance or DONOT inform it right.
10:17 hdl      So it should be automatically generated.
10:17 hdl      Maybe.
10:17 thd      hdl: many fields could be set automatically to help the lazy.  Or provide a screen that the user cannot move past without setting some required information to prompt the lazy :)
10:21 thd      hdl: The character set could certainly be guessed at by the system from what was in the record for setting automatically.  If the user wanted to use another one than the default the user would need to change something.
10:23 thd      hdl, owen, osmoze: conversion to 8859 for the OPAC would still be needed for the user systems outside the library that do not have UNIMARC support installed.
10:26 thd      hdl, owen, osmoze: I suspect that no one wants to refuse them service when their system shows C e?z a n n e instead of CĂ©zanne.
10:27 owen     Not exactly refusal of service, but of course that would be ideal
10:28 thd      or whatever actually happens, the link to I ? Unicode was :-)
10:32 thd      hdl: so the plugin that you referred to, does that create a new item record based on the parent record?
10:33 hdl      thd : I agree with you : Conversion to ISO-8859 for OPAC.
10:34 hdl      thd :No. it is a plugin to help biblio input.
10:34 hdl      unimarc_field_4XX are UNIMARC links with other biblios.
10:35 thd      hdl: what does it do to help biblio input and what is the name of the plugin?
10:36 hdl      So you can link this plugin to a 7XX field and click on the "..."
10:36 hdl      You are promtpted do choose a biblio then it fills all the field subfields with correct values and adds a $9 with biblionumber.
10:37 hdl      find it in : value_builder/unimarc_field_4XX.pl
10:40 thd      hdl: Does it create a new child or sibling record based on the parent record, as I had been discussing with owen a month ago?   I pasted that previous text into yesterday's discussion.
10:43 hdl      It creates the field and fills in the subfield.
10:43 hdl      It creates No new MARC::Record.
10:45 thd      hdl: Owen had been looking for a simple way to add a serial record for an issue of a serial, a month ago, where the cataloguer would not need to retype everything, especially as that issue would be discarded after a brief time and record would then be deleted.
10:47 thd      hdl: So value_builder/unimarc_field_4XX.pl merely creates the linking field with the appropriate information to the linked record.
10:50 hdl      thd : yes : is this not enough ?
10:50 hdl      Is this not what you expected ?
10:54 thd      hdl: There is still typing the title, subjects, etc. everything that goes into creating a record if all you want to do is merely copy a record and link back to it as a parent record.  Then issue information for a serial and a separate barcode might be added to the child record and nothing more need be typed.
10:58 thd      hdl: With reference to the issue that kados had brought up yesterday, where the library is retaining all the issues and not discarding them.  Your serials management procedure does not allow for separate barcodes for each issue does it?
11:00 hdl      Wait Wait Wait.
11:01 hdl      "My" 4XX link management was not meant only for serials.
11:01 hdl      It was for all 4XX management.
11:01 thd      hdl: I realise that.
11:02 hdl      And as such, you cannot expect it to do the job.
11:02 thd      :)
11:03 hdl      I said yesterday that serial issues management should be workde upon depending on what the library wants.
11:04 hdl      Wether serial or issues are important is a criterium to decide how to manage issues.
11:04 thd      hdl: So your existing serials management procedure would not provide for a separate barcodes for each issue?
11:05 hdl      ATM : It just create an entry in serial tables.
11:05 hdl      It is not MARC compliant.
11:05 hdl      And do not link to an item.
11:06 hdl      Most of the time, in library, you cannot issue an issue....
11:06 hdl      funny : issue an issue :))))
11:06 owen     It depends on the library, obviously: we circulate individual issues all the time.
11:08 thd      hdl: That level of MARC compliance requires careful scrutiny just to apprehend the documentation.
11:10 hdl      owen : i know that some library consider issues as a book and circulate it.
11:10 hdl      But you have then to fill in a biblio for each issue.
11:11 hdl      Indeed, quite often, the persons who contributed to the issue change.
11:11 hdl      Subjects change to.
11:13 hdl      So it would be difficult to make it simple or automatic IMHO
11:13 owen     But with a garden-variety publication like National Geographic, for instance, the information stays the same each month.
11:13 hdl      Sure
11:14 hdl      But one must be aware of the worst case.
11:14 hdl      which comes quite oftne, when going deeper and deeper.
11:16 hdl      Maybe a quick biblio could be done, which the librarian could get and improve when he has time.
11:16 hdl      Could be a workaround.
11:17 hdl      But maybe not THAT MARC compliant.
11:17 owen     Why not simply offer the option of adding an item record for each newly received issue?
11:18 owen     Of course since Koha can't reserve by item, you still couldn't place reservations on individual issues...
11:22 hdl      The problem is that IMO, it is not just an item record, but a new biblio each time...
11:24 hdl      If you make a resarch on High-Availability Articles you should get all the issues related to this subject, and not only the name of the serial.
11:25 thd      hdl: Putting the true equivalent of your serial holdings table in UNIMARC would entail the UNIMARC's holding format as a separate record.  Although, I imagine that you might spell out the table in 995 $u.
11:26 hdl      Sorry.
11:28 thd      hdl: Could libraries not use UINIMARC 995 $u to designate an issue?
11:29 hdl      Wouldnot it be AWFULL :(
11:29 thd      hdl: a particular issue of a periodical, or a particular bound volume of issues for a period of time in UNIMARC 995 $u.
11:30 thd      hdl: Yes, the UNIMARC holdings format with holdings in a separate record is better for that purpose.
11:31 hdl      When you have been subscribing to a serial for 50 years... Do you htink One MARC record wil contain everything ?
11:32 thd      hdl: Well, the holdings format has a way of compressing the information but if there are 50 years worth of issue based barcodes that may become too much for record size limits.
11:33 hdl      ;)
11:33 hdl      And It would not be satisfying.
11:33 thd      hdl: kados had an example of a library with over 300 issues of a periodical each with its own barcode.
11:34 hdl      members donot look for a particuliar issue, but most of the time for a piece of information, for an article...
11:34 hdl      thd : All in ONE MARC record ???
11:38 thd      hdl: I had thought that the usual case was that a user consulted a serials indexing and abstracting service to find the article citation with volume and issue number and then the catalogue to find the general serial record.
11:40 hdl      thd.. Maybe you are right. Unfortunately, I have never been a "library rat".
11:40 thd      hdl: Of course that process is much faster with OpenURL links to a full text database or the catalogue from within an indexing and abstracting service.
11:41 thd      hdl: I grew up in libraries.  Or never grew up in libraries :-)
11:42 hdl      Lucky one :)
11:45 thd      hdl: Most general libraries that I know do not separately barcode each issue for long term retention on an issue basis.  Issues are gathered and bound into volumes to reduce the individual items over time.
11:46 owen     I suspect it's going to take a paying customer to further develop serials in any meaningful way.  It's too abstract right now.
11:46 thd      hdl: kados has found some other types of serial management practices that approach the MARC record size limits.
11:51 hdl      thd : I saw this practise of "Voluming" issues.
11:51 hdl      Well, folks, sorry... but... It is time.
11:51 thd      hdl, owen: How might the OPAC show the main serial parent record and not show 300 child records as well if the child records were all separate bibliographic records?  The separate bibliographic records would exist to not exceed the maximum MARC record size when issues were recorded individual with their own barcode on a permanent basis.
11:52 hdl      It is my sister birthday.
11:52 owen     Wish her a happy birthday from Ohio for me :)
11:52 thd      happy birthday hdl's sister
11:52 hdl      YESS...
11:52 hdl      She'll be very pleased.
11:53 hdl      But Indeed, serials ARE a CRUX we should lean on to understand precisely What is asked.
11:53 hdl      I'll reead the logs.
11:54 owen     thd: have you had experience with serials in other ILSes?
11:55 thd      owen: I have had little experience with other Ila's outside the OPAC and disembodied cataloguing modules.
11:56 owen     That's part of our problem.  Neither have I--and part of the reason for that is that vendors like to put a hefty price tag on serials support
11:56 thd      s/Ila's/ILS's/