Time Nick Message 10:24 slef yeah, it's none of those AFAICT 10:23 owen http://filext.com/detaillist.php?extdetail=LEX 10:23 slef so I'm guessing binary and not nice 10:23 slef but trying to less it just ate my session :-/ 10:23 slef it's some sort of book data 10:21 slef Has anyone seen .lex files before? 04:31 chris yep, i have to be up at 7am to see kickoff, so im going to sleep too 04:30 russ see you tomorrow morning 04:26 alaurin hi all, yes, it's a big day today, for a lot of french 03:46 chris big game today eh arnaud? 03:46 chris hi all 02:59 hdl hello world 01:59 ToinS hello world 20:40 thd kados: why would Encode::decode be needed to pass the content of a string if there is no need to parse the actual characters in the string? 20:38 thd kados: what differences do you find between browsers if anything other than display? 20:36 kados thd: I'm not entirely convinced it's a browser problem 20:29 thd Burgunavia: the ordinary version of reality viewed through Debian GNU/Linux etch with more font and encoding problems than if I were using some system running Internet Explorer. Yet I do not want any of the Internet Explorer OS's or Internet Explorer even. 20:26 Burgundavia thd: what do you use? 20:25 thd I want a different universe or at least a different OS 20:24 thd N 20:04 thd chis: do you understand when and where Encode::decode should be needed and when it would be unnecessary? 19:53 thd kados: do I understand correctly that Perl cannot manipulate the byte string correctly unless Encode::decode has been used? 19:45 thd tumer: must have been introduced with Klingon :) 19:44 tumer thd: these are Unicode terms i am not very familiar with 19:42 thd tumer: what does the 'N' represent? 19:41 tumer thd:from now on i use NFC for composed and NFD for decomposed characters 19:41 tumer composed character format 19:40 tumer s/Record/USMARC 19:40 thd tumer: what is NFC? 19:40 tumer and all my templates do break with the new Record.pm but mine 19:39 tumer kados:still all my records are NFC'ed both in zebra and in Mysql 19:39 thd tumer: so if it is only in Zebra then there is no code to display the record detail or support author searching by clicking on the author link? 19:38 tumer thd:this record seems to exist only in zebra 19:37 thd kados: that is the record found by searching for yasa 19:36 tumer thd:copying and pasting you can search but not from keyboard 19:36 kados thd: I've no idea 19:36 thd kados: why is there nothing displayed for this record outside of the result set? http://zoomopac.liblime.com/cgi-bin/koha/opac-detail.pl?bib=99999999 19:35 tumer then it should work 19:35 kados yep 19:35 tumer is it NFC'ing the results? 19:34 kados tumer: quite even 19:34 kados tumer: it's quit short :-) 19:34 thd kados: yet my posted author search fails, why? 19:33 tumer i have to read entity_encode 19:33 tumer i have to check your _entity_encode 19:33 tumer we put them in zebra composed 19:32 thd kados: we can find them in most cases by copying the string and pasting the query so they must be stored correctly. 19:32 tumer our keyboards produce composed characters 19:32 kados or just while they are being pulled out for display? 19:32 kados so do we entity encode them before we import? 19:32 kados ok ... 19:32 tumer otherwise you will not be able to serach them from the keyboard 19:31 kados tumer: should do the trick 19:31 kados tumer: _entity_encode() subroutine 19:31 kados tumer: i have code that does that in Record.pm in HEAD 19:31 tumer yes 19:31 kados tumer: does that mean we should entity encode them? 19:31 tumer some browsers like IE can handle them ddecomposed 19:31 tumer kados:XML definition and web character definition requires them to be composed 19:30 kados tumer: (ie, are they stored wrong in the db, or just being mangled on the way out?) 19:30 kados tumer: or is it just the display? 19:30 thd kados: are you not referring to a mere browser problem if it is browser specific? 19:30 kados tumer: so is it necessary to import using your modified USMARC.pm? 19:29 tumer its a matter of firefox and safari beeing poorer in displaying decomposed characters 19:29 tumer kados:i know and in IE they will. If you use my corrected version of USMARC.pm thay all will look correct 19:28 kados tumer: in safari the O in that page works, but the s doesn't ... in firefox, neither one does :( 19:28 thd tumer: http://zoomopac.liblime.com/cgi-bin/koha/opac-zoomsearch.pl?query_form=ccl&op=get_results&ccl_query=O%CC%88rneklerle 19:27 tumer thd:? 19:27 thd tumer: if everything is correct why does my posted author search finding nothing and failing to use the correct characters 19:27 chris ill have to google 19:27 chris good question 19:27 kados chris: setting aside for the moment how to fix the databases of already existing Koha users 19:26 kados chris: so what do I need to do to create all the Koha tables and columns correctly? 19:25 thd kados: yes if you create the table new it is not a problem but then you do not have your data 19:25 chris yeah that would be best joshua 19:25 tumer kados:your page looks correct on IE 19:25 chris i suspect out convet_utf8.pl needs to be fixed 19:25 kados then import the data 19:25 kados though it would probably be better to create them as utf8 19:24 chris hmmm i suspect so 19:24 kados so we need to convert them then, eh? 19:24 kados right 19:24 chris but all the columns are still latin1 19:24 tumer kados: if you had accented characters in your template you will see that they break as well 19:24 thd tumer: http://zoomopac.liblime.com/cgi-bin/koha/opac-zoomsearch.pl?query_form=ccl&op=get_results&ccl_query=O%CC%88rneklerle 19:23 chris i think what we have done is told the tabel to be utf8 19:23 chris yeah how to fix it 19:23 chris ahh so i suspect that there is the problem eh 19:23 kados chris: now, the trick is, what's the solution? :-) 19:23 kados chris: just emailed you what the mysqldump looks like 19:23 thd tumer: I find that the canonisation rules are not supporting Altug\x{0306} Osman, 19:22 kados chris: you're right about the problem 19:22 dewey tumer: i'm not following you... 19:22 tumer dewey:upgrade zebra 19:22 kados chris: you there? 19:22 tumer thd:ok this resolves the problem of query displaying correctly 19:22 dewey the problem is probably if there is no data in the sort field zebra throws and error and does not even return the unsorted results 19:22 chris whats the problem? 19:21 chris not me 19:20 tumer thd: just checking 19:20 thd tumer: you can see the result of what kados had done with encode earlier 19:18 tumer well so you are ahead of me 19:18 thd tumer: kados says that most of my suggestions are for version 4 19:18 thd :) 19:17 tumer working on version 3.5! 19:17 tumer i am currently deeply modifying intranet 19:16 thd :0 19:16 tumer its just a line of code which currently is too much 19:16 tumer thd: currently yes 19:15 thd tumer: you had not added this to the record detail display it is only in the multi-record result set 19:15 bob cool 19:15 chris ahh 19:14 tumer bob: i am cleaning those. left owers of previous anymation we tried 19:14 bob this line in the code though looks wrong -> <script language="JavaScript1.2" fptype="dynamicanimation" src="file:///C:/Program%20Files/Microsoft%20Office/Office10/fpclass/animate.js"> 19:14 chris very cool 19:13 thd rach: click on a call number in an OPAC search result set to see the best Koha demo ever :) 19:13 tumer end of october it will take you to the correct face of the shelf- i hope 19:13 bob woo that is fancy 19:12 tumer thanks guys 19:12 thd fantastic it works 19:12 chris thats cool as hell tumer 19:12 chris sweet works now 19:11 tumer correct server now i hope 19:10 thd tumer: you never modified the record detail display to support the map? 19:09 thd :) 19:08 chris :) 19:08 tumer oops sorry wrong server 19:08 chris im still seeing location tumer 19:06 tumer any luck anyone? 19:06 tumer ii have renamed it to locationonly 19:05 tumer bob:good point 19:04 bob eg window.location.href 19:04 chris hmm that might be it 19:04 bob i know there one can call window.location 19:04 bob is location a reserved word? 19:04 chris but i might be missing something 19:04 chris is one is resizable and the other isnt 19:03 chris the only difference i can see 19:03 chris and for some reason that bit doesnt work .. the help function above it .. works just fine 19:03 chris yep 19:03 tumer } 19:03 tumer help_window.focus(); 19:03 tumer var help_window = window.open('/cgi-bin/koha/opac-location.pl?location=' + locationc +'&lib='+lib,'_blanck','height=300,width=640,status=no,toolbar=no,menubar=no,location=no,resizable=no,top=0,left=0,scrollbars=1'); 19:03 tumer function location(locationc,lib) { 19:03 tumer and then.. 19:02 tumer it just calls href="javascript: location('<!-- TMPL_VAR NAME="classification" -->','<!-- TMPL_VAR name="location-only" -->');"> 19:02 chris the help one seems to work, the location one doesnt 19:02 chris } 19:02 chris help_window.focus(); 19:02 chris var help_window = window.open('/cgi-bin/koha/opac-location.pl?location=' + locationc +'&lib='+lib,'_blanck','height=300,width=640,status=no,toolbar=no,menubar=no,location=no,resizable=no,top=0,left=0,scrollbars=1'); 19:02 chris function location(locationc,lib) { 19:02 chris help_window.focus(); 19:02 chris var help_window = window.open('/cgi-bin/koha/opac_help.pl?help_template=' + help_topic ,'_blanck','height=400,width=400,status=no,toolbar=yes,menubar=no,location=no,resizable=yes,top=0,left=0,scrollbars=1'); 19:02 chris function help(help_topic) { 19:01 bob says it may not load or link to file:///C:/Program%20Files/Microsoft%20Office/Office10/fpclass/animate.js 19:01 chris the help popup works 19:00 bob me reloads 19:00 chris i think thats why location isnt a function 19:00 chris uncaught exception bit bob? 19:00 chris on page load 18:59 chris yeah and before that do you get the error 18:59 bob says location is not a function 18:59 thd tumer: are you using a variable 'General Collection', just the sort of name with a space that bob cautioned about? 18:59 chris its the javascript its complaining about 18:58 chris line 232 it seems to be moaning about 18:57 tumer thd: i delete all the names the templates have in them as well otherwise it does not popup in IE 18:57 thd tumer: do any of your names have holes in them? 18:56 chris take a look at that page bob, thats where it happens 18:56 chris http://library.neu.edu.tr/cgi-bin/koha/opac-search.pl?op=do_search&type=opac&keyword=fish&ttype=exact&op=do_search 18:56 bob where name was like 'koha popup' 18:56 bob eg window.open(url,name,...) 18:55 thd tumer: you would never do that would you? 18:55 bob which firefox can handle but IE could not 18:55 bob i found the problems with most javascript popups that work in firefox but not IE were due to people trying to give the popup window a name with a space in it 18:54 tumer chris:i will look into that. its a simple localion.pl Perl script supposed to run in a window 18:53 chris uncaught security exception 18:53 chris it doesnt like the location function 18:52 tumer chris:does it say what error is that 18:52 thd tumer: Is there not one set of JavaScript code to make the same function work for any browser? 18:52 tumer same javascript supposedly! 18:51 chris i get a javascript error in firefox 18:51 tumer thd: ihave to check that and see what i can do.Firefox popups do not popup in IE so i change them 18:50 thd tumer: I cannot get the map to pop-up on Firefox and I do not have Opera installed at the moment because Opera prevented me from upgrading x-windows libraries. 18:50 tumer thd: can you see the little popup plan? 18:49 thd tumer: kados and I still have extra hair 18:49 tumer thd:is there a bug i only look at it in IE? 18:48 tumer chris:i have checked characters hex by hex since this problem started used to have white hair now no hair at all 18:47 thd tumer: does the JavaScript for the call number map pop-up only work on Internet Explorer? 18:47 chris i was just checking that i was viewing it in utf8 by toggling off and on 18:46 chris yes thats right tumer 18:46 tumer chris we are supposed to look and see in utf8 dont we? 18:46 chris cant hurt 18:45 tumer so i say we bug the Marc::Record maintainer and see what he says 18:45 chris but if im viewing the page as utf8 it looks fine 18:45 chris if i change encoding from utf-8 to anythign else i do 18:45 chris nope 18:44 tumer changing the language to turkish from the top of the page do you see any anomalies 18:44 chris its looking ok to me 18:44 tumer library.neu.edu.tr uses this fix 18:43 tumer i am currently using the fix i reported in production 18:42 tumer Encoding the whole Output.pm does not work either 18:42 tumer but i already have utf8 characters on the template and they get broken 18:41 thd tumer: It did make some difference in the display 18:41 chris right 18:41 tumer what that does is to create the problem of query 18:41 tumer chris:what kados is proposing of using Encode for the query dows not work ( i tried them all) 18:40 thd tumer: any call number? how complete is the map? 18:40 tumer click on the LC call number 18:40 rach thanks very much 18:39 thd tumer: yes, what records bring up the animated walking path to the book? 18:39 tumer rach:library.neu.edu.tr should be working 18:39 thd chris: only pleasant screaming of course :) 18:38 chris :-) 18:38 tumer well sleepless nights help koha:-) 18:38 rach howdy 18:38 chris :-) dont remind me 18:38 chris thank you 18:38 thd chris: and in December the screaming starts 18:38 tumer chris congragulations 18:37 chris anytime in the next week or so, the kicking should start apparently :) 18:37 tumer so is he/she started kicking yet? 18:37 chris yep, december 19th 18:36 tumer hi chris btw i hear wife expecting a baby? 18:36 chris we had a public library go live on saturday, so ive been pretty busy 18:36 thd tumer: he had another suggested solution today using Encode but he found problems persisted and I was uncertain why that code would solve anything. 18:36 chris hopefully in the next few days 18:36 chris tumer: im working on getting some space to test it 18:35 tumer well i have posted it to devel expecting people to use and comment on it 18:34 thd tumer: so if he does not have your fix installed he cannot test it :) 18:33 tumer thd:well i think he did not still try it so i cannot see the difference on his server 18:32 thd tumer: where does kados find that the problem still exists which you do not see? 18:31 tumer thd:i believe i have fixed it but kados does not 18:31 tumer that is what is causing all this havoc of utf8 display problems 18:31 thd tumer: I mean what is wrong that you had not fixed? 18:30 thd tumer: what is wrong with the current version of MARC::Record? 18:29 thd bob: yes are you using a set of authority records or are you just creating a few token values in an array? 18:29 tumer as long as you dont need MARC8 to UTF8 conversion you will see that all problems go away 18:28 tumer thd:regarding utf8 i believe we have a bigger problem than anticipated. Revert to earliear version of MARC::Record 18:28 bob but not sure what i'm doing is right 18:28 bob and playing with them 18:28 bob hmm, i looking at the files in the /value_builder/ sub dir 18:27 thd bob: the best version is rel_2_2 just before paul's recent commits to fix a bug reverted things somewhat 18:27 tumer bob:thats how we fill tag 650 with data from somehere else 18:26 tumer bob:what u are asking i belive is using authorities not value builder. 18:25 tumer hi thd 18:24 bob 2.2.5 i think 18:24 thd bob: what version of the record editor are you using? 18:20 thd tumer: the scary UTF8 monsters dragged kados away. 18:12 thd tumer: have you come to save us from the scary UTF8 monsters? 18:11 thd bob: the current value builder design which expects one value for all the subfields in the field works best for names in 1XX, 7XX, etc. 18:09 thd s/most cases/most cases with real authority records from real authority files/ 18:09 bob ta 18:08 thd bob: this is a problem especially for subjects, where the value builder does not work as well as it might for filling subdivided subject headings in most cases.. 18:07 bob right 18:06 thd bob: I think in future, multiple value builders will function so that the free floating subdivisions $z, $x, $y, and $v can be assigned using the value builder in context separately from $a if needed. 18:02 thd bob: so you should be able to fill one 650 after another with different values and if you have a good version of the record editor you can fill all the subfields in the field at once and not just $a. 18:01 bob umm not sure 18:00 thd bob: so you mean the repeatable 650 of which the first subfield is $a and $a as we know for this case is non-repeatable while the 650 is repeatable? 17:59 thd bob: I answered above about repeatable fields such as repeatable MARC 21 field 650 for topical subjects 17:58 bob i'm meaning tag 650 subfield a 17:58 bob ah right 17:57 thd bob: so one confusion I had was how you were naming fields or subfields in asking your question. 17:55 bob if you are using repeatable fields 17:55 bob will the popup be repopulated with the right info 17:55 bob but i'm wondering if you then edit the record using the popup 17:54 bob i understand the popups can take a bunch of info and place it in right places in the marc record form 17:54 thd bob: I had difficulty understanding your question as asked on koha-devel 17:53 bob yep 17:53 thd bob: I answered above for the record editor. Is your context the record editor? 17:52 bob yep 17:52 bob ah ype 17:52 bob so if you have a bunch of repeatable fields for 650a with info in them will clicking the ... grap the info from all or just the one 17:52 thd bob: had you asked a question on koha-devel to day relating to value builder pop-up? 17:51 thd bob: the value builder is often filling repeatable fields and filling them individually. 17:50 thd hdl: I thought tumer had reported that it was not a problem for him using Internet Explorer set to UTF8 preferred but only for the rest of us using Firefox or whatever else. 17:48 bob eg when editing a record does the popup get info from just one field or many? 17:47 bob i've a question about the value_builder plugins do they work for repeatable fields? 17:47 hdl Normally, the display is UTF-8 encoded (sent by apache.) And even with MY environment full UTF-8, it is ugly displayed. 17:45 thd hdl: are you saying that I ought to see the single UTF8 character in the record editor even if my environment is not UTF8? 17:44 thd hdl: I cannot save a test record on your test system to see that the end result is fine because hidden MARC tag for the record ID is not being automatically filled in the default framework. 17:43 hdl thd: IMHO, you should see the accented character and not "as a composed single character glyph but I see the correct characters for each of the two UTF8 bytes which will form the character." 17:40 thd hdl: everything looks correct to me except that on my ISO 8859-1 environment system I do not see the accented character represented as a composed single character glyph but I see the correct characters for each of the two UTF8 bytes which will form the character. 17:32 chris interesting, i just mysqldumped some, and loaded them back in .. and they appear ok ... ill check in with you tomorrow joshua and see what we are doing different 17:16 chris then take a look at that file 17:16 chris select * from tablename into outfile '/tmp/somefile'; 17:15 chris in mysql 17:15 chris one thing to try would be in 17:12 chris ppl all over the world are using it to do just that 17:12 chris no it can 17:11 kados maybe that mysql can't handle charsets :-) 17:11 chris there will be something we are missing 17:11 chris cool 17:11 kados yes 17:11 chris this data was loaded in after the table was set to utf8 ? 17:10 chris you are aware changing a table to utf8 doesnt change the data in the table eh? 17:10 kados I'll see if I can lose some more hair on this tomorrow :-) 17:09 kados thanks for the help 17:09 kados well ... I've got to get going 17:09 chris yeah thats its default 17:09 kados I also tried --default-character-set=utf8 17:09 kados it just mangles it in a different way :-) 17:09 chris and still mangled? 17:08 kados I've tried that too :-) 17:08 chris and just take a look what it looks like 17:08 chris when you mysql dump it 17:08 chris --default-character-set=latin1 17:07 kados en_US.UTF-8 17:07 kados l$ echo $LANG 17:07 kados btw: 17:07 chris i have a theory, can you try this for me please 17:07 chris ok 17:07 kados I woudln't think it would since 4.0 didn't support _any_ character sets :-) 17:07 chris yeah 17:06 kados in mysqldump? 17:06 chris -compatible=mysql40 make any difference? 17:03 kados debian sarge's : mysql Ver 14.7 Distrib 4.1.11, for pc-linux-gnu (i386) 17:03 chris what version of mysql do you have? 17:03 chris how odd 17:02 kados :( 17:02 kados there is no diff between > and -r 17:02 kados diff -q kohawipo.utf8.sql kohawipo.utf8.sql2 17:01 chris hmm 16:59 kados because I can read raw MARC with it 16:59 kados which I know does utf8 ok 16:59 kados mangled in vi 16:58 chris or some editor we know parses utf8 ok 16:56 chris or if you open it in jed? 16:56 kados I'll give it a shot 16:56 chris is it mangled if you import it back in 16:55 chris i wouldnt expect that to work 16:55 kados chris: I tried the -r option, but when I do 'cat koha.sql |more' I still see mangled characters 16:51 chris no problem ive done it with maori data and its worked .. so hopefully it should fix the problems you are having with import/export too 16:50 hdl chris : thx too 16:50 kados chris: thx for the tips, I'll try that 16:50 hdl and select the first accented value. 16:49 hdl try and click on the ... on field 632 then search egypt* 16:48 thd hdl: I am uncertain just what I should do to observe this problem 16:47 chris to load it back in 16:47 chris mysql -e 'source /patch-to-backup/backup-file.sql' 16:45 hdl and you will be able to see a problem when creating a new Biblio and getting accented authorities 16:44 hdl thd: Try on i3.hdlaurent.paulpoulain.com 16:44 thd chris: I see, I missed your -r option posting while I was typing 16:44 chris use -r instead 16:44 chris im saying that the redirect to a file doesnt work with utf8 16:43 chris there arent 16:42 thd chris: but I see no UTF8 characters in your command 16:41 thd s/UTF(/UTF8/ 16:41 chris MySQL>source dbdump.sql 16:41 chris Example: 16:41 chris Using MySQL, source the sql script. 16:41 chris >mysqldump -r dbdump.sql 16:41 chris Example: 16:41 chris Using MySQLDump, make sure to use the '-r <filename>' option, instead of redirecting command-line output to a file ('> <filename>'). Redirection does not capture UTF characters properly, while the -r option does. 16:41 thd chris: when would you eve need to use UTF( on the command line? 16:40 thd kados: you did not seem to have any shortage of hair to me 16:40 chris that wont be a big thing, it'll be the commandline client not supporting utf-8 16:40 thd hdl: do you have an example of the problem and an example of what the 'blind detail' page is itself with and without encoding problems? 16:39 kados it makes me question the chosen platform ! 16:38 kados accented characters outside the ascii range 16:38 kados characters are mangled 16:38 kados and then import them 16:38 kados if you attempt to mysqldump your utf8 records out 16:38 kados with the rel_2_2 version 16:38 kados another encoding problem has turned up 16:31 kados hmmm 16:30 hdl kados : it is the page that fills the bibliographic datas with the Authority input. 16:28 kados what is blind-detail-biblio? 16:28 kados interesting 16:27 kados hdl: yes? 15:51 hdl if I choose an accented term for 700 or 650 in a bibliographic record from an authority (say égypte for instance), it wil result into a strange encoding when blind-detail-biblio occurs 15:49 hdl taht is, 15:49 hdl I checked his problem and has the same on 226 version for utf-8 management 15:48 hdl kados