IRC log for #koha, 2015-11-09

All times shown according to UTC.

Time S Nick Message
00:33 dcook joined #koha
00:41 ex-parrot joined #koha
00:41 eythian ex-parrot: welkommen
00:41 ex-parrot hey eythian
00:41 wahanui go back to bed, eythian
00:41 ex-parrot I just came to harass cdickinson
00:42 eythian seems reasonable
00:42 ex-parrot (koha is cool too)
00:42 dcook ^
00:51 Francesca I don't think cdickinson is around
00:52 ex-parrot I managed to get ahold of him :)
00:52 cdickinson I just got back, Francesca
00:52 Francesca sup
00:53 Francesca_ joined #koha
00:54 Francesca_ ugh mac just crashed
01:01 jamesb joined #koha
01:11 liz joined #koha
01:33 eythian what's the table that stores what permissions borrowers have?
01:35 eythian oh
01:35 dcook hehe
01:35 eythian maybe user_permissions
01:35 eythian or maybe not
01:36 eythian it's hard to tell
01:36 dcook Well, flags in the borrower table, and then I think a mix of..
01:36 eythian ah
01:36 dcook `permissions` i think is subpermissions?
01:36 eythian I bet it's the flags one I want
01:36 dcook I would think so
01:36 dcook And then you compare that against...
01:36 dcook `userflags`
01:37 dcook Right, `userflags` has the permissions... `permissions` has the subpermissions...
01:37 dcook In the best naming scheme ever
01:37 eythian ikr
01:37 dcook And I think `user_permissions` might store somethihng about subpermissions?
01:38 eythian all I need to see is the users that have any permissions, so it's easy enough.
01:38 eythian huh, it's weird seeing the name of someone I used to work with pop up in there.
01:38 eythian oh, she used to work there too
01:38 eythian I'd forgotten about that
01:39 eythian ("there" being the client whose database I'm poking in)
01:39 eythian[…]s-of-li%C3%A9bana <-- unrelated
01:39 dcook And people think there isn't any interesting data in library databases
01:39 dcook The potential for data misuse seems fairly high to me...
01:39 dcook hehe
01:39 dcook I like that image
01:40 eythian that's why we don't let people get to the database who aren't supposed to :)
01:42 eythian oddly, I'm totally not finding the user I'm expecting to
01:43 eythian maybe it's vanished
01:43 dcook deletedborrowers?
01:43 eythian nah, I expected it on staging, it's probably been zapped by a database refresh
02:18 indradg joined #koha
03:08 dcook eythian: I was just thinking... in theory Zebra could handle other formats than just MARC, yeah?
03:08 dcook The problem would be with Koha
03:08 eythian well
03:08 dcook That it would try to get MARC out and that wouldn't work for non-MARC
03:08 eythian Probably
03:08 eythian but, I don't know if you can mix them up
03:09 eythian however, it's way outside my scope of knowledge of zebra.
03:09 dcook hehe
03:09 dcook Fair enough
03:09 dcook Yeah, it's outside mine as well
03:10 dcook Thinking about all the different challenges we face with supporting different metadata formats...
03:11 dcook Zebra is a free, fast, friendly information management system. It can index records in XML/SGML, MARC,
03:11 dcook e-mail archives and many other formats, and quickly find them using a combination of boolean searching
03:11 dcook and relevance ranking.
03:11 wahanui relevance ranking is broken by QueryAutoTruncate
03:11 dcook Shh
03:11 * dcook wonders if he was the one who wrote that..
03:12 dcook But even if you stored multiple things in Zebra... you'd need some sort of way to get them out..
03:12 eythian yeah
03:12 eythian Koha really expects marc formats
03:12 dcook Yeah
03:12 dcook I wonder how Zebra returns thing...
03:13 dcook Does it store the original and its own internal record?
03:13 dcook It certainly stores the latter..
03:13 dcook If only windows would let me open things..
03:13 eythian heh windows
03:13 eythian I have no idea how zebra stores things
03:13 dcook Kill windows with fire..
03:13 ex-parrot left #koha
03:14 eythian not ex-parrot, windows
03:14 eythian zebra is mostly a mysterious black box to me
03:14 dcook Fair enough
03:17 dcook Come on, Zebra, tell me your secrets...
03:19 kathryn joined #koha
03:19 dcook This install looks funny..
03:28 dcook Ahhhh
03:28 dcook "A parser for binary MARC records based on the ISO2709 library standard is provided, it transforms these
03:28 dcook to the internal MARCXML DOM representation"
03:28 dcook Actually, save the ahhh moment for later..
03:32 dcook I think I understand everything except how it outputs usmarc..
03:32 dcook iso2709 rather..
03:38 dcook It doesn't store iso2709 internally... not even in GRS1
03:38 dcook GRS1 used it's own internal format..
03:40 dcook But we only care about DOM now..
04:01 dcook So with DOM... it'll parse iso2709 into MARCXML... and it just stores MARCXML as MARCXML in the Zebra storage...
04:01 dcook Which is why the identity.xsl will return MARCXML for "elements marc" and "elements marcxml"
04:01 dcook Neato burrito
04:03 dcook In theory you could index whatever into Zebra and get out MARCXML... so long as you could specify a XSLT to make it so
04:04 dcook Although the smarter thing to do would probably be to define some sort of intermediary format...
04:04 dcook Which would require re-doing all the XSLTs for course
04:05 dcook Or would it..
04:06 dcook You could actually expand the existing detail and result XSLTs...
04:06 dcook Or refactor them to make the better...
04:06 dcook We could have templates for discrete parts of the detail page
04:07 dcook Title template, author, the rest, etc.
04:07 dcook Well... that could be tricky..
04:07 dcook Might get confusing for developers
04:08 eythian that's sooorrrrttttaaa what I'm doing with ES
04:08 eythian though, it's not really used
04:08 eythian it could be though
04:08 dcook How do you mean?
04:09 eythian all the fields are converted (mappings) to things like "title" and so on
04:09 eythian it's more aimed at searching though
04:09 eythian I don't think putting display elements into the index is a good idea
04:09 eythian I think putting the record in and converting it to display after fetching is better
04:09 dcook Agreed
04:10 eythian after all, you might want different displays in different situations
04:10 dcook The XSLT for the detail page could check if it's MARCXML or whatever else
04:10 dcook Yep
04:10 eythian I think you would need an XSLT for each format you support
04:10 eythian it's the only sane way
04:10 dcook It would be the easier way :p
04:10 eythian insofar as XSLT is sane
04:10 dcook True that
04:10 dcook You could do multiple ones..
04:11 dcook And have one top level one
04:11 dcook Well...what do I mean..
04:11 dcook Really one XSLT
04:11 eythian I suppose you could
04:11 dcook You'd import the other ones in just to get access to their templates
04:11 dcook The idea would be to make the search results and detail pages modular
04:12 dcook So the XSLT needs to create certain blocks of HTML
04:12 dcook Title, author, details, etc
04:12 dcook And the way that each metadata format does that is up to its particular template
04:12 dcook Which the top-level template will call depending on the metadata format
04:12 dcook Easy peasy
04:12 eythian yeah true
04:12 eythian good thinking
04:12 dcook Thanks
04:13 eythian you could even have the XSLT extract parts and put them into XML, then that's rendered using a display XSLT
04:13 eythian so you'll have a conversion XSLT for each format, but only one display one for each display situation.
04:13 eythian e.g. results vs. detail vs. staff client
04:14 eythian Alien vs Predator vs Brown vs The Board of Education
04:14 dcook I'm not sure I entirely follow but I think I agree
04:14 dcook Ah, yes, I getcha
04:14 dcook That's another way of doing it
04:14 dcook I think it would be more work in the short-term, but much smarter in the long term
04:15 eythian little modular lego bricks :)
04:15 dcook And you can specify that extraction XSLT in the Zebra config
04:15 dcook Have an intermediate format that you can use for the display XSLT
04:15 eythian can you do it per record?
04:15 dcook per record type
04:15 dcook So yes
04:15 eythian ah OK
04:15 dcook So I think search could be achievable
04:16 dcook And in the short-term... I'm thinking that search is the most important thing
04:16 dcook Because that leads you to external resources or to items
04:16 dcook Items...
04:16 wahanui i guess items is library-speak for books.
04:16 dcook I suppose items is the next hurdle
04:17 dcook (And facets for search)
04:17 dcook Although the Zebra facets would get around that problem..
04:17 dcook And you could even do facets the way we already do them and hard code them for different formats... which is ugly but doable
04:17 dcook Just need to detect the namespace in the incoming record which is easy enough as well
04:17 dcook Ah but not if we do the intermediary format
04:18 dcook But then if we do the intermediary format, we could use a standard way of getting our own facets out
04:18 dcook Bam!
04:18 dcook Logic all over the place..
04:19 dcook I suppose any multi-format effort would have to look at search... and item creation... possibly all cataloguing...
04:19 dcook Well maybe not cataloguing. That could maybe be done later. As we could get the records in...
04:19 * dcook is just babbling now
04:21 Francesca joined #koha
04:24 dcook I wonder how it would work with RDF...
04:24 dcook I suppose you can display RDF however you want
04:24 dcook In terms of linked data, you'd just need to publish the original record at the prescribed URI
04:24 dcook (and probably via a SPARQL endpoint)
04:25 dcook And that's something that magnuse is already working on
04:25 dcook was/is*
04:25 dcook I think..
04:30 Amit_Gupta joined #koha
04:36 dcook Although I guess with RDF it's much more difficult..
04:39 dcook As sometimes you need to follow the link..
04:53 Amit_Gupta hi dcook
04:53 dcook hey Amit
04:57 Francesca hey dcook
04:57 dcook hey Francesca
04:57 wahanui is waging a war against the video driver that continues to fail on her vm
04:57 Francesca lol
04:58 Francesca not anymore
04:58 Francesca how goes aussie
04:58 dcook Mmm it goes
04:58 dcook Reading more about RDF :p
04:59 dcook You?
04:59 wahanui You are welcome
05:06 Francesca apart from being welcome, I am kinda bored
05:06 Francesca need a cat
05:06 Francesca cats?
05:06 wahanui The only good cat is a stir-fried cat.
05:07 dcook Bored?
05:07 dcook Cats will certainly keep you from being bored
05:08 Francesca my cats are outside right now
05:08 Francesca they dont want to cuddle me or entertain me
05:09 dcook You could figure out how to integrate RDF into Koha :p
05:20 Francesca what is RDF
05:20 wahanui RDF is, like, just spinning round in circles on the water
05:20 Francesca apart from that
05:20 dcook resource description framework
05:21 Francesca and what does it do
05:21 dcook Example:
05:21 dcook Good question :p
05:21 dcook In theory, it lets you link machine readable resources together
05:21 Francesca ah
05:21 dcook So look at that web page
05:22 Francesca sorry don't know if I can help with that - interesting but I don't quite understand
05:22 dcook I don't know if I understand quite yet either :p
05:22 Francesca lol
05:22 dcook The idea is that you have lots of datasets out there, and instead of copying the data... you put in a pointer
05:22 dcook So for "birthPlace", you put a link to London rather than saying London
05:23 Francesca huh
05:23 dcook Well, let's use you for an example
05:23 dcook Instead of using the plain text of your parents names in your record
05:23 dcook You'd put in the URIs to their records
05:23 Francesca ok
05:24 Francesca so kinda just adding all the info in one place via links
05:24 dcook So if we're trying to view a record on you, the server should parse your record, and realize that your parents are referred to at those URLs
05:24 dcook Yep
05:24 dcook So the server would fetch the records for your parents and display their info
05:24 dcook How it decides what info to show... I don't know
05:24 dcook I think that's up to the server... which is where it all falls apart imho
05:24 Francesca lol
05:26 Francesca server side I might not be as helpful
05:27 dcook Well, you could do it via AJAX :p
05:27 dcook I just figure your machine has to be programmed with some knowledge of the schema of the remote resource
05:28 dcook I think there might actually be some facility for that with <rdfs:label/> in RDF/XML...
05:28 dcook Which makes sense to me
05:28 Francesca heh I've never worked in AJAX
05:28 dcook If there is a RDF field to refer to the thing in plain text at its most base level
05:28 dcook Well jeez, Francesca!
05:29 Francesca lol
05:29 Francesca my work mainly consists of looking at something and going hey can I re-style it
05:30 dcook That's also good :)
05:31 Francesca got any webpages that need styling
05:31 Francesca I could do that quite quickly
05:31 dcook Oh probably but not off the top of my head
05:32 Francesca lol if you think of any you know where to find me
05:33 dcook Sounds good ;)
05:35 Francesca good luck with the RDF
05:36 dcook Thanks
05:36 dcook At the moment, I don't know how it could ever be useful for searching
05:37 dcook Browsing for sure
05:37 dcook If you have the record retrieved, it could show you all sorts of interesting links
05:37 dcook But how do you retrieve that record?
05:37 Francesca yeah
05:37 Francesca an interesting problem for sure
05:37 dcook For sure
05:38 dcook I suppose you could query your RDF triple store and say who who has a reference to this URL?
05:38 dcook But that's not very user friendly
05:39 Francesca yeah
05:51 dcook @later tell magnuse Yo, we should chat RDF!
05:51 huginn dcook: The operation succeeded.
05:53 Francesca dcook: if its any comfort I think my job searching is going about as well as the RDF
05:53 dcook :(
05:53 dcook That doesn't sound good
05:53 dcook You're still in uni, yeah?
05:56 Francesca yeah
05:56 Francesca I was hoping to get an internship this year but things didnt work out
05:57 dcook :(
05:59 Francesca so now I have to find a summer job because work would be good
06:32 saiful joined #koha
07:07 laurence joined #koha
07:14 * magnuse waves
07:17 magnuse dcook: yeah, let's talk rdf at some point
07:20 magnuse dcook: have you seen[…]i/Linked_Data_RFC - it's a braindump that is getting to be a few years old, but i think it still sums up my thinking on rdf and koha
07:21 cait joined #koha
07:22 rocio1 joined #koha
07:24 sophie_m joined #koha
07:33 marcelr joined #koha
07:33 marcelr hi #koha
07:34 cait morning all
07:34 cait bbl
07:34 cait left #koha
07:36 magnuse hiya marcelr and cait
07:36 fridolin joined #koha
07:36 magnuse ...and matts and fridolin
07:36 matts hi !
07:36 fridolin bonjour magnuse and all of u
07:37 magnuse @wuner boo
07:37 huginn magnuse: downloading the Perl source
07:37 magnuse @wunder boo
07:37 huginn magnuse: The current temperature in Bodo, Norway is 8.0°C (8:20 AM CET on November 09, 2015). Conditions: Mostly Cloudy. Humidity: 66%. Dew Point: 2.0°C. Windchill: 2.0°C. Pressure: 29.15 in 987 hPa (Falling).
07:47 reiveune joined #koha
07:49 reiveune hello
07:50 magnuse bonjour reiveune
07:57 alex_a joined #koha
07:57 alex_a bonjour
08:00 fridolin left #koha
08:05 Francesca joined #koha
08:12 jajm hi
08:15 Francesca hey jajm
08:17 cait joined #koha
08:17 * cait waves
08:17 * Francesca waves back at cait
08:17 cait hi Francesca :)
08:18 Francesca sup :)
08:18 cait did you have a nice weekend?
08:28 fridolin joined #koha
08:33 gaetan_B joined #koha
08:33 gaetan_B hello
08:33 wahanui hello, gaetan_B
08:34 indradg joined #koha
08:41 Francesca joined #koha
08:42 Francesca cait: yes, good weekend thanks
08:48 Amit_Gupta heya gaetan_B
08:48 wahanui i think gaetan_B is working at Biblibre and did the nice new start page together with asaurat or a fan of icons
08:50 gaetan_B hi Amit_Gupta
09:00 Traumatan joined #koha
09:03 wilfrid joined #koha
09:17 rocio1 left #koha
09:19 paul_p joined #koha
09:34 fridolin I have a problem with debian jessie + indexdata repo, libnet-z3950-zoom-perl can not be installed
09:34 fridolin because of a dependancy on perlapi-5.18.2, but the version 5.20.0 is installed
09:34 fridolin maybe  a problem with the packaging by indexdata
09:47 Joubu hi #koha
09:47 cait hi Joubu!
10:09 mveron joined #koha
10:09 mveron Hi #koha
10:14 Joubu cait: Are you able to recreate the failure on the report tests?
10:15 Joubu I have just recreated a fresh DB and the tests still does not fail for me
10:15 cait only the one time
10:16 cait not sure what needs to be done to get it back to the state before...
10:16 cait it would point to something changing in the database then?
10:16 cait i can try and load an older dump - but only tonihgt
10:17 cait i was testing with my 'play' db - so ther are probably some reports
10:20 Joubu Tomas told us that Jenkins loads a new DB now, so no reports should be there
10:20 cait hm eah
10:20 cait some variable not set the first time?
10:20 cait I am really not sure why this happens
10:48 Joubu @later tell tcohen please send me a dump of a DB to recreate the failure on Repots_Guided.t
10:48 huginn Joubu: The operation succeeded.
11:25 Traumatan joined #koha
11:28 Amit_Gupta hi cait
11:52 cait first time patch writer.... someone around for testing? bug 15136
11:52 huginn Bug[…]_bug.cgi?id=15136 enhancement, P5 - low, ---, contact, Needs Signoff , Display item's homebranch in patron's fines list
12:08 clrh joined #koha
12:09 tcohen joined #koha
12:11 drojf joined #koha
12:11 drojf hi #koha
12:18 tcohen morning #koha
12:19 tcohen Joubu: it is in /home/jenkins/koha_3_20_00.sql.gz
12:19 Joubu ok thank
12:26 tcohen Joubu: the problem I think is on side effects from previously run tets
12:26 Joubu tcohen: the tests pass with this DB...
12:28 tcohen Joubu: :-(
12:29 tcohen maybe try to run the previously run tests (on jenkins list)
12:30 Joubu tcohen: how do you generate the 'fresh' DB on jenkins?
12:30 tcohen Joubu: only the file I mentioned earlier
12:30 tcohen (3.20.0 +
12:31 Joubu tcohen: I got some errors on updatedatabse: DBD::mysql::db do failed: Table 'audio_alerts' already exists [for Statement "
12:31 Joubu and some otheres
12:31 Joubu so it's not a 3.20 db
12:31 tcohen http://jenkins.koha-community.[…]etedBuild/console
12:31 Joubu hum...
12:32 Joubu ok I know, I haven't erase the previous DB, so the tables existed
12:32 tcohen that sounds promising :-D
12:33 * tcohen goes prepare coffee while the kohadevbox fires
12:33 * cait waves
12:33 Joubu ok drop + create + update + prove => ok
12:33 tcohen Joubu: f*ck
12:34 tcohen on jenkins, the only one failing is GetTopIssues.t, maybe is date-related?
12:34 tcohen jajm: were you involved on GetTopIssues.t?
12:34 Joubu no
12:35 Joubu ha sorry :)
12:35 jajm tcohen, yes i think...
12:36 tcohen it's been failing for the last couple weeks
12:36 tcohen if you have the time, i'd appreciate that you took a look
12:37 jajm tcohen, where can I find the test output ?
12:37 tcohen http://jenkins.koha-community.[…]er_D7/572/console
12:37 jajm thx
12:39 Joubu jajm: at one point, the AI for biblio and biblioitems are not in sync anymore
12:39 jajm 17:33:39 [10:54:22] t/db_dependent/Circulation/GetTopIssues.t ................. ok     9952 ms
12:39 jajm tcohen, ^
12:39 Joubu http://jenkins.koha-community.[…]etedBuild/console
12:39 Joubu 18:28:12 t/db_dependent/Circulation/GetTopIssues.t               (Wstat: 65280 Tests: 0 Failed: 0)
12:41 paul_p joined #koha
12:41 tcohen ah, wrong one
12:41 tcohen heh
12:41 cait tcohen: could you foward the latest release notes to my workmail maybe?
12:41 tcohen I think it is TooMany.t's fault
12:42 tcohen I'm regenerating them anyway
12:42 tcohen jajm: http://jenkins.koha-community.[…]etedBuild/console
12:42 tcohen it's a problem with the tests
12:43 tcohen and I had the test failing without biblionumber-bibioitemnumber divergence
12:44 tcohen the problem is the biblioitemnumber retrieved is NULL
12:44 tcohen don't ask me how that happens :-D
12:46 tcohen jajm: to reproduce what i do is
12:47 tcohen vagrant up jessie ; vagrant ssh jessie ; cat /vagrant/scp koha_3_20_00.sql.gz | sudo koha-mysql kohadev
12:47 tcohen sudo koha-shell kohadev; cd kohaclone ; perl installer/data/mysql/
12:48 tcohen and then if you run prove t/db_dependent/Circulation/GetTopIssues.t it passes
12:48 tcohen (run it many times, it works)
12:48 tcohen then run prove t/db_dependent/Circulation/TooMany.t prove t/db_dependent/Circulation/GetTopIssues.t => FAIL
12:51 meliss joined #koha
12:54 * magnuse waves again
12:55 JoshB joined #koha
12:56 tcohen jajm: ok?
12:57 drojf joined #koha
13:16 tcohen jajm:
13:18 tcohen @wunder cordoba, argentina
13:18 huginn tcohen: The current temperature in Cordoba, Argentina is 29.0°C (10:00 AM ART on November 09, 2015). Conditions: Scattered Clouds. Humidity: 48%. Dew Point: 17.0°C. Pressure: 29.80 in 1009 hPa (Steady).
13:22 magnuse @wunder boo
13:22 huginn magnuse: The current temperature in Bodo, Norway is 8.0°C (1:50 PM CET on November 09, 2015). Conditions: Mostly Cloudy. Humidity: 62%. Dew Point: 1.0°C. Windchill: 4.0°C. Pressure: 29.09 in 985 hPa (Steady).
13:22 magnuse tcohen wins
13:23 tcohen heh
13:30 jajm tcohen, i reproduce the bug, investigating now... :)
13:30 tcohen jajm: AWESOME
13:38 Dyrcona joined #koha
13:39 talljoy joined #koha
13:40 jajm tcohen, is there a bz where i can send a patch or should I create a new one ?
13:41 tcohen there's a bug i think
13:41 tcohen Joubu: ?
13:42 tcohen jajm: fill a new one
13:44 jajm ok
13:45 Joubu I didn't open it
13:48 NateC joined #koha
13:48 jajm tcohen,[…]_bug.cgi?id=15158
13:48 huginn Bug 15158: minor, P5 - low, ---, julian.maurice, Needs Signoff , t/db_dependent/Circulation/GetTopIssues.t is failing in Jenkins
13:49 tcohen jajm: i owe you a pastis bottle this time
13:49 mario joined #koha
13:49 mario morning
13:49 jajm tcohen, thanks, but i hate pastis... :)
13:49 tcohen ok, so i'll keep the bottle :-P
13:53 cait heh
14:02 Dyrcona joined #koha
14:03 nick joined #koha
14:03 pastebot "tcohen" at pasted "Joubu: the test" (25 lines) at
14:03 nick joined #koha
14:04 wnickc joined #koha
14:04 Joubu tcohen: I suspect that it is mysql specific
14:05 tcohen yeah, but that's what the DBIC people told on
14:05 tcohen s/on/me/
14:07 tcohen Joubu: the other option is to just create two consecutive biblios and check they have consecutive biblionumbers
14:07 tcohen what do u think #koha
14:07 Joubu this is certainly better
14:13 pastebot "tcohen" at pasted "Joubu: it even reads better :-P" (21 lines) at
14:16 fridolin left #koha
14:20 fridolin joined #koha
14:21 tcohen bug 15159
14:21 huginn Bug[…]_bug.cgi?id=15159 normal, P5 - low, ---, tomascohen, Needs Signoff , TestBuilder behaviour on AI values should be tested
14:24 tcohen jajm++
14:27 cma joined #koha
14:34 saiful joined #koha
14:51 JoshB joined #koha
14:53 amyk joined #koha
15:02 paul_p joined #koha
15:05 cait jajm: around? i had a line with pull downs showing up when i had applised the statistic wizard cataloguing patch
15:05 cait no label in front
15:06 cait it's still there with the patch applied
15:06 cait the second row
15:07 cait i have no idea what it does, the code refers to it as cotedigits.... ?
15:11 gaetan_B in the cirulation rules the "Default checkout limit by patron category" says "For this library, you can edit rules for given itemtypes, regardless of the patron's category."
15:11 gaetan_B but http://schema.koha-community.o[…]r_circ_rules.html doesn't store the branch
15:12 gaetan_B is it stored somewhere else or is the interface wrong?
15:13 Traumatan joined #koha
15:13 gaetan_B ok this is indeed stored elsewhere
15:17 fridolin joined #koha
15:21 xarragon Hmm, I was changing and tried running ./t/db_dependent/Members_Attributes.t which failed. It fails for current master as well as 3.20.3. Am I doing anythign wrong?
15:21 xarragon 1/60 Can't call method "default_privacy" on an undefined value at /var/koha/Koha/C4/ line 750
15:24 xarragon Guess I need to check the test invocation, b/c the file is only 350 lines long
15:24 gaetan_B cait: "cote" is the french word for "callnumber" maybe a bad translation from jajm ?
15:25 xarragon Ah, I see.. probably need to have PWd be t subdir
15:25 jajm cait, I don't have this row... :/
15:27 tcohen xarragon: it is not failing in master for me
15:27 jajm cait, ah yes I see... :)
15:27 tcohen xarragon: are you sure you updated your db, etc?
15:28 xarragon tcohen: I think I just invoked it wrong. Just checked the wiki.
15:29 xarragon tcohen: I jsut called prove ./t/db_dependant/Members_Attributes.t
15:30 cait jajm: did you find it? was making tea :)
15:31 jajm cait, i don't know exactly what it's supposed to do, but it's in master too, maybe you should file a new bug for this ?
15:32 cait yeah that's what i meant in my comment - old problem.. but gah. :)
15:32 cait not going to force you to fix it ;)
15:33 jajm cait, it's really nice of you :)
15:34 jajm it's there since 2005 ...
15:34 cait one of the old and weird things we have where probably noone knows what it's supposed to do... )
15:52 bdonnahue joined #koha
15:53 huginn New commit(s) kohagit: Bug 15158: Fix t/db_dependent/Circulation/GetTopIssues.t <[…]84f0d3502ae3153bd> / Bug 14867: userid not generated when defined in BorrowerUnwantedField <[…]ad09c17fb7b7b8913> / Bug 14388: Funds should be sorted by budget_code <
16:03 TGoat joined #koha
16:06 tcohen @later tell marcelr can we fix that "WARNING: You do not have upload_path..." thing? setting a default path and taking care of the installer creating it?
16:06 huginn tcohen: The operation succeeded.
16:10 rocio joined #koha
16:11 francharb joined #koha
16:38 laurence left #koha
16:46 pianohacker joined #koha
16:46 pianohacker hello
16:51 saiful joined #koha
16:53 cait hi pianohacker
16:53 fridolin left #koha
17:03 drojf joined #koha
17:11 gaetan_B bye
17:17 reiveune bye
17:17 reiveune left #koha
17:18 cait left #koha
17:47 wnickc @later tell wizzyrea I redid bug 14739 for 3.18, give it a once over and let me know if you spot any problems
17:47 huginn wnickc: The operation succeeded.
18:05 huginn New commit(s) kohagit: Bug 14402: (QA followup) Add notes to usage text about --fees <[…]90409de1d30a0ddbd> / Bug 14402: Make purge_zero_balance_fees() delete fees with NULL balance. <[…]c181bba8d55c43bc4> / Bug 14402: Add option --fees to /misc/cronjobs/ <http:/
18:21 francharb joined #koha
18:32 BobB joined #koha
18:47 huginn New commit(s) kohagit: Bug 8064: DBRev <[…]08c8f9e4ac80be5f3> / Bug 8064: Fix unit tests for createMergeHash <[…]264537cb1bec3ef54> / Bug 8064: Little fix for 003, 005, 008 in MARC21 <[…]a53560b411dd44e14
19:07 huginn New commit(s) kohagit: Bug 15036: Do not overwrite complete status in basket ops <[…]600015c9be3320bf0>
19:10 magnuse joined #koha
19:31 tcohen vagatn destroy jessie
19:31 tcohen oops
19:31 tcohen bye #koha
20:22 sophie_m left #koha
20:42 liz joined #koha
20:51 cdickinson joined #koha
20:52 bdonnahue hey guys is there a way to have koha import mark 21 records in batch?
20:58 magnuse joined #koha
21:11 pianohacker bdonnahue: yes, there's the stage marc records for import under Tools, there's the script
21:18 bdonnahue pianohacker: thanks
21:18 bdonnahue is there any way to have koha "look up" a marc record from a reputable source like lib of congress and then import it if it is correct?
21:21 rocio_away left #koha
21:58 eythian bdonnahue: yes, that's Z39.50 under the cataloguing screen
22:16 eythian dcook: your email feels a bit stream-of-consciousness :)
22:16 dcook eythian: Well... you know me
22:16 dcook Definitely stream-of-consciousness
22:16 dcook I really did intend it just to be about library size at first
22:16 dcook :p
22:16 eythian heh
22:17 eythian it sure veers off
22:17 dcook I figure no one is really going to read the whole thing anyway
22:17 dcook Yeah, the size thing had to do with database size and I started thinking about how we probably could handle a lot more records than we do
22:17 dcook I mean... I work on other databases with millions upon millions of rows and it's not a big deal
22:18 dcook But it's more complex than that of course
22:18 dcook eythian: I wrote a thing for library school students about how it would be wise to concentrate more on learning than marks
22:19 dcook Closed it with </unsolicited rant>
22:19 dcook Which a friend pointed out could be my middle name :p
22:19 eythian haha
22:19 dcook Whoops. Wrote eythian in the 40$c
22:19 dcook Don't think you're the transcribing agency..
22:19 pianohacker ha!
22:19 eythian I am now!
22:20 dcook hehe
22:40 burdsjm__ joined #koha

| Channels | #koha index | Today | | Search | Google Search | Plain-Text | plain, newest first | summary