Time  Nick       Message
10:43 thd        slef: z39.50 servers are only good.  You should let them all in.  My firewall does not block z39.50 responses over port 80 but my security is low :)
10:41 slef       thd: I'm 99% sure the firewall is blocking z39.50 to all other servers :-/
10:41 slef       thd: :)
10:41 slef       thd: no and no
09:51 thd        paul: are you still here?
09:30 thd        paul: are you there?
08:53 kados      (btw: entity encoding the utf8 may be the answer to our previous encoding probs, I thought of that last night ...)
08:52 kados      thd: can't wait to hear about it ;-)
08:51 thd        kados:OK tumer and I have found something
08:51 kados      thd: I'll be back later
08:51 kados      thd: I've got to head out for a couple hours
08:50 kados      thd: I will be soon :-)
08:50 thd        slef: do you have a valid reliable Z39.50 target configured and are you searching for a record that can certainly be found?
08:48 thd        tumer: you also have the other problem with not clearing $9.
08:47 thd        tumer: I hope that he has not gone away
08:47 thd        tumer: I have been trying to show him this problem since late yesterday
08:46 tumer      thd:if you catch kados can you please mention him this coversation?
08:45 thd        :)
08:45 tumer      thd:unfortunately i have to go to a reception now, but at least we are starting to all see the same problem which is good
08:44 thd        tumer: has your record been saved with the record editor or only imported?
08:44 thd        tumer: that is very bad
08:43 tumer      thd: no but i have the same authorname i mean
08:43 thd        tumer: do you mean that you added it to your system?
08:42 tumer      thd:the same record with the same search will yield the double encoding on my system and kados's which both use zebra
08:42 thd        kados: are you awake yet?
08:41 thd        tumer: I expect that the production system where I have seen double encoding is not using Zebra but I am not certain
08:40 tumer      thd:we may be sitting on another problem that this thing actually happens with ZEBRA only and not with SQL?
08:40 thd        tumer: I have s[pent most days recently too busy to advance my installation of Koha
08:39 thd        tumer: I have not taken the time actually to set up Zebra yet.
08:38 thd        tumer: my search results are coming from SQL
08:38 tumer      thd: are you using zebra?
08:38 thd        tumer: i expect that the double encoding I see on the production system where I have been editing records is from some part of the system not recognising that the record was already converted from MARC-8 to UTF-8.  There had somewhat recently been a problem for MARC::File ::XML doing that.
08:37 tumer      thd: by the way is your serach results coming from ZEBRA or mysql?
08:36 tumer      thd:now this is serious. What you are saying is that its not a double encding problem but a problem of X-windows or Windows?
08:35 thd        tumer: I do not know the solution for double encoding other than finding where it is happening and stopping it.
08:34 thd        tumer: I have not seen the double encoding problem on the installation on my system at home but I have seen it on a production system that I have been working on for kados
08:34 tumer      thd:so you mean you know the solution or this?
08:33 thd        tumer: my X-windows may also be partly misconfigured like everyone's
08:33 tumer      doing the same search on my system will give the double encoding problem
08:32 tumer      thd:yes the problem is how did you mabage to get this page display correctly like this and not have the double utf8 problem?
08:31 thd        tumer: yet, because different fonts are use on the page with an individual record instead of multiple search results the individual record looks fine
08:30 thd        tumer: thank you now I know that I suffer deeply for my freedom I see some very strange out of order accent on this page
08:30 tumer      thd: by the way this is what i am trying to achieve
08:29 tumer      thd: only in name Valérie there is an accent
08:29 thd        tumer: so the problem is with X-windows or fonts on my GNU/Linux system then.
08:28 slef       thd: yes, and I can run it through the debugger.  It forks but never changes from Still ?? requests to go.
08:28 tumer      thd:i see accented e that is é
08:27 thd        tumer: what accented characters do you see?
08:27 tumer      thd:no it looks perfectly fine
08:27 thd        tumer: do you see a difference between the representation of the accented character in the results for search sting and author results column?
08:25 tumer      thd: the page looks normal with french accented characters
08:21 tumer      thd: what do you want me to do?
08:20 tumer      thd: i am back now
08:15 thd        tumer we could test whether you can see the same problem when you are off phone
08:11 thd        slef: can you determine that the daemon is running with the ps command?
08:10 thd        slef: do you mean the connection is working fine or that you have no error messages? :)
08:09 thd        tumer: he has only been back from ALA yesterday
08:09 thd        tumer: kados did say that he would hunt down the double encoding problem that you reported
08:08 slef       thd: none, as far as I can tell.
08:06 thd        tumer: I am observing this character display problem especially in the uncommon cases where a fixed width font is used but maybe it is my GNU/Linux system
08:03 thd        ?
08:03 thd        slef: what problems does you Z39.50 connection report
08:02 thd        s/post/port/
08:02 thd        tumer: my problem of posting UTF-8 content is gone now with Firefox 1.5.04.  What I am identifying is a problem for character display that may not affect Turkish characters but certainly exists for French character display in both Firefox and the Opera post from Windows.
08:00 tumer      thd:as long as i keep with true type (unicode) characters of windows i do not et that problem
07:59 thd        tumer: this problem is browser independent
07:59 tumer      thd:yes i know but not for internet explorer which i use
07:58 thd        tumer: there is still also a problem with the fonts used in the CSS for every template which will obscure the actual character content for some UTF-8 multibyte characters.
07:56 slef       Is there any easy way to debug a z39.50 connection that I think is being blocked by a firewall, or do we need to wait for the administrator to interrogate the firewall?
07:56 tumer      thd:well at least that releives me some of mine
07:55 thd        tumer: kados is extremely fatigued by encoding problems
07:54 tumer      kados have been silent about encoding problems i reported anyone knows why?
07:54 thd        tumer: I have seen more examples of double encoding UTF-8 to UTF-8
07:53 thd        oh, I suspected that all the bugs would not be gone
07:52 tumer      thd:i already have that
07:52 thd        25/06/06 12:09:18+-5	<kados:#koha>	my $record=MARC::Record->new_from_xml($xml, 'UTF-8');
07:52 thd        25/06/06 12:09:17+-5	<kados:#koha>	and make sure they look like this:
07:52 thd        25/06/06 12:09:11+-5	<kados:#koha>	look for all instances of 'new_from_xml' in addbiblio.pl
07:52 tumer      hdl:the changes in marc editor breaks authorities
07:52 thd        25/06/06 12:09:00+-5	<kados:#koha>	for a temporary fix
07:52 hdl        you know : when using $field->subfield('a') it returns either a string OR a list depending on the variable on the right of your =
07:51 thd        tumer: I have found the fix kados suggested on #koha
07:51 tumer      hdl:the problem i am having is not of multiple subfields but of completely wrong fields getting cleared or even sometimes filled
07:50 hdl        It should be a problem of PERL context.
07:50 tumer      even if you prepare multiple x's on the editor they get filled with same data allover.
07:49 hdl        I could investigate but not before next week.
07:49 tumer      hdl:well there is. It does not
07:48 hdl        I have not tested. But paul told me that there could possibly be a problem.
07:47 tumer      hdl:say a subject authority has 150$a 150$x and another 150$x does it get transferred to the marc editor correctly?
07:46 hdl        Yes, I understand.
07:45 tumer      well thats not so easy of not upgrading as the system is now half-breed
07:44 thd        tumer: kados made a fix which he has not posted yet but which he described the needed changes
07:43 thd        tumer: that version has a problem when cloning fields and certainly you should not upgrade the addbiblio.pl
07:43 tumer      i am afraid to upgrade now
07:42 tumer      my marc editor is version the one just before paul broke
07:42 thd        tumer: which version are you using?
07:41 tumer      the problem occurs if you use this cloning of subfieds and then use yhe authorities
07:41 tumer      hdl:the main problem is not cleaning
07:41 thd        tumer: are you using a recent version from CVS?
07:40 thd        oops $9 I mean
07:40 tumer      hdl:yes on the screen buth the marc record could be left with $9 authid filled
07:40 hdl        (thd: $9 kohaauth number)
07:39 hdl        But on my devweek version, which is quite old, the "clearing" of authorities seems to work.
07:39 thd        tumer: so the problem is that $3 is not being cleared when the popup is opened?
07:38 tumer      hdl:its the same way with npl templates. You use a popup for athorities
07:38 hdl        It's ok.
07:38 tumer      the last line is a mess of mistakes.sorry
07:37 tumer      its just that i write to td more so used to write ths automatically
07:37 thd        I realise you pinged me by mistake
07:36 tumer      hdl:the previous line was supposed to be for you not to thd
07:36 thd        hello tumer
07:36 tumer      oh hi thd
07:35 thd        yes, I was about to guess that
07:35 tumer      thd: thats exacly the what i am talking about
07:35 hdl        tumer: It is not so easy to understand but with default templates, you have to open the popup and clear entry.
07:35 tumer      thd:i am talking about using authorities in normal marc editor
07:34 thd        tumer: I had thought that there had been no changes to the authorities editor in a very long time
07:34 tumer      the blind-biblio-search.pl only cleans subfields a..z
07:33 tumer      when you clean a field of authority the authority number does not get cleaned..
07:33 hdl        Is there an official branch for devweek merged with rel_2_2  or is it that You did a merge ?
07:32 tumer      hdl:also i think there is a missing code
07:32 hdl        wow, what a mess.
07:31 tumer      i am using npl templates though
07:31 thd        tumer: yes
07:30 tumer      thd: i am using dev_week merged to rel_2_
07:30 tumer      juts have a page where there are more than one field that uses authorities in te same page
07:29 tumer      if you want to reproduce the problem..
07:29 hdl        devweek ?
07:29 hdl        which version are you working on ?
07:28 hdl        BEST is the fiend of good. :(
07:28 tumer      we are in the middle of cataloguing 10,000 books and the system is now broken
07:27 tumer      i think so because we never had it before
07:27 hdl        Is this only because of the latest changes in MARC editor ?
07:26 tumer      so authorities is now btoken with this new editor and i had to stop using authorities
07:26 hdl        Should be a problem of dupping the authtagtoreport using the good order.
07:26 hdl        Yes.
07:26 tumer      hdl: do you understand the problem?
07:25 hdl        No, but I have VERY simple authorized forms.
07:25 tumer      hdl:did you get this?
07:24 tumer       it deletes a different tag
07:24 tumer       or if you try to delete tem..
07:24 hdl        OUPPS.
07:24 tumer      they get filled into differnt tags ..
07:24 hdl        Oh Yes.
07:24 tumer      and then tries to field the field from the authorities..
07:23 tumer      if the cataloger plays with cloning subfields or changing their order..
07:23 hdl        an url ?
07:22 tumer      the biblio-blind-search.pl gets confused
07:22 tumer      hdl: on the editor when you get an authority with all these repeatable fields..
07:22 hdl        Can you tell me ?
07:22 hdl        If I can help.
07:21 btoumi     hi all
07:21 tumer      i am having probleems with authorities
07:21 hdl        how are you ?
07:21 hdl        hi
07:21 tumer      hi hdl?
07:20 dewey      tumer is here for a few seconds ;-)
07:20 hdl        tumer ???
07:20 hdl        I was having lunch and overlooked your beep
07:20 hdl        tumer[A]: I'm here.
06:40 tumer      hdl:are you around?
04:00 ToinS      hi slef
03:39 slef       hii
02:36 btoumi     hello hdl
02:36 ToinS      salut hdl
02:36 ToinS      très bien
02:36 hdl        hello world
02:35 btoumi     ?
02:35 btoumi     ca va
02:34 ToinS      salut bruno
02:29 btoumi     hi toin's
02:16 ToinS      hello
02:16 btoumi     hi all
02:11 osmoze     hello
21:41 russ       thanks for all of that, that link to the loc was helpful, i have found what i need.
21:31 thd        russ: $e, itself, is not repeatable
21:29 thd        russ: $e also contains what might have been in a repeated $c which is allowed but I have not noticed it used.
21:28 thd        russ: $e contains everything for the accompanying atlas which might have been in a repeated $b were it allowed.
21:27 thd        russ: 300 ##$a271 p. :$bill. ;$c21 cm. +$eatlas (37 p., 19 leaves of plates : 19 col. maps ; 37 cm.)
21:26 thd        russ: consider the example of the atlas in http://www.loc.gov/marc/bibliographic/ecbdphys.html#mrcb300
21:26 russ       ok cool thanks
21:24 thd        put all the information in $e itself for the accompanying material
21:24 thd        russ: If you were wanting to use a repeated $b to add additional information about accompanying material in $e, which is not allowed then ..
21:22 thd        russ: I may guess at what you want to know for encoding 300
21:22 thd        russ: http://www.loc.gov/marc has the most up to date information for the concise format information
21:20 russ       what resource do you use to find this out?
21:20 russ       thanks
21:20 russ       right ok
21:20 thd        russ: 300 $b is not repeatable
21:19 thd        russ: 300 is repeatable even 300 $a is repeatable and $a is seldom repeatable
21:18 thd        russ: apparently, I was wrong
21:16 thd        yes, I believe
21:16 russ       or should it be repeatable?
21:15 russ       the 300 b subfield, is that repeatable?
21:15 thd        yes
21:15 russ       i have a marc question for you if you have a minute
21:15 russ       hiya
21:13 thd        yes russ
21:13 russ       thd are you about?
21:13 dewey      thd was last seen on #koha 3 hours, 57 minutes and 37 seconds ago, saying: kados: are you still there? [Fri Jun 30 10:15:30 2006]
21:13 russ       dewey seen thd?
17:15 thd        kados: are you still there?
15:44 kados      :-)
15:38 thd        kados: I guess that a benchmark is not needed in instances when you have a 1000 times difference.  No one cares if it is 998 or 1002 times too long for that particular case.
15:35 kados      officially :-)
15:35 kados      I haven't benchmarked anything yet
15:33 thd        s/my/me/
15:33 thd        kados: my spell checker sometimes gives my strange transformations that I do not always catch
15:31 thd        s/bookmarked/benchmarked add field, and other transformations/
15:30 kados      bookmarked?
15:29 thd        kados: have you bookmarked MARC::Record relative to MARC::File::XML?
15:27 thd        kados: yet the browser only has to manage the browser view
15:26 kados      a 10K MARC21 file could easily be 100K in MARCXML
15:26 kados      of course, XML is much more verbose :-)
15:26 kados      not that I know if
15:25 thd        kados: is there any performance difference for Zebra presenting one format or other for search results?
15:24 kados      but currently, I'm using xml
15:24 kados      there is code to do both
15:23 thd        kados: what format does the record editor save to the database for Zebra?
15:23 kados      yes
15:23 thd        kados: is importation 1000 times slower if the records have already been converted to UTF-8?
15:20 thd        kados: what format does the record editor save to the database for Zebra?
15:19 kados      though I haven't tested
15:19 kados      I think so
15:19 thd        kados: can we break the MARC record size limit without problem using MARC-XML once the record has been added or newly created?
15:17 thd        kados: does the record editor also submit the resulting record from edits int the same format?
15:16 thd        kados: does the record editor add a new record in MARC-XML or MARC?
15:15 kados      about 1000 times slower :-)
15:15 kados      yes
15:13 thd        kados: importation is much slower if it were in XML?
15:12 kados      internally it's stored in Zebra's internal format
15:12 kados      and MARCXML after that
15:12 kados      we are using binary MARC for the initial import
15:12 kados      not quite
15:12 thd        kados: we are using MARC instead of MARC-XML in Zebra for performance reasons?
15:11 kados      thd: I'm all ears :-)
15:11 thd        kados: I remembered an important question
15:10 thd        ahh
15:10 kados      thd: jmf@kados.org is the right address I think
15:10 kados      thd: I'm here
15:03 thd        ?
15:03 thd        kados: how can I assign a bug to you
15:01 thd        kados: when I have tried assigning or CC a bug to you it tells me that jmf@liblime.com is not a recognised email address for that purpose.
14:59 thd        kados: bug filed but I have not been able to assign it to you
13:58 kados      yep
13:58 thd        kados: editing fixed fields without some extra aid is crazy though
13:57 thd        kados: you do not have to use the plugin and so there is a workaround
13:57 kados      IMO anything that prevents us from creating valid MARC21 is a blocker  :-)
13:56 kados      blocker means that you can't release the software until it's fixed
13:56 thd        kados: is that what blocker means?
13:55 kados      thd: could you file a bug in bugzilla ... mark it as 'blocker' since it prevents creation of valid MARC21 records
13:51 GrahamDoel thank you kados and owen,  problem solved.  Thanks again
13:46 thd        kados: maybe this happens for the 008 plugin under other conditions but this was the condition where I noticed it.
13:46 kados      should be a simple fix
13:45 kados      yes
13:45 thd        kados: is that clear?
13:45 thd        kados: if the first 6 positions for record creation date are blank then using the plugin does not set them automatically but eliminates the first six positions moving everything over by 6 positions
13:42 kados      thd: do tell
13:42 thd        kados: there is a bug in the 008 plugin
13:32 GrahamDoel ahh.. I see.  No connection at...
13:28 GrahamDoel yes, I found that on a list somewhere and did check that I had done.
13:27 owen       GrahamDoel: you filled in a search term, right? An ISBN or title?
13:25 GrahamDoel ok, thanks... I'll look now
13:25 kados      GrahamDoel: whereever you specified in the z3950-daemon-options file
13:24 GrahamDoel ok... could you remind me where they are?
13:24 kados      you need to see what the log says to be sure
13:24 kados      GrahamDoel: check the logs
13:24 owen       kados: can you assume the daemon is running properly if the script doesn't generate any errors?
13:23 owen       Hmmm... that's one I /don't/ know.
13:23 kados      :-)
13:23 thd        /msg kados kados:  I am still finishing the Afognak job
13:22 GrahamDoel I have run the start daemon script, but don't know how to check
13:22 kados      thd: yep
13:22 GrahamDoel when I do an add biblio enter the isbn and click the z3950 search it just returns z3950 Still ?? requests to go
13:22 thd        kados: are you back?
13:21 owen       What are the symptoms of your problem? Are the searches not returning any results?
13:20 GrahamDoel thanks
13:20 owen       I can try
13:20 GrahamDoel ?
13:20 GrahamDoel I have a fresh install of koha and am strugglind with the z39.5 searches, do you think you might be able to point me in the right direction
13:19 GrahamDoel Hi owen
13:19 owen       Hi GrahamDoel
12:30 kados      ahh
12:24 slef       (buggy RFC-ignorant mailserver)
12:24 slef       no it won't, as it's bouncing to the From not the envelope
12:06 kados      mailman will do it automatically
12:03 slef       ok np