16:00:44 #startmeeting RELENG (2019-07-18) 16:00:44 Meeting started Wed Jul 17 16:00:44 2019 UTC. 16:00:44 This meeting is logged and archived in a public location. 16:00:44 The chair is mboddu. Information about MeetBot at http://wiki.debian.org/MeetBot. 16:00:44 Useful Commands: #action #agreed #halp #info #idea #link #topic. 16:00:44 The meeting name has been set to 'releng_(2019-07-18)' 16:00:44 #meetingname releng 16:00:44 The meeting name has been set to 'releng' 16:00:45 #chair nirik tyll sharkcz masta pbrobinson puiterwijk mboddu dustymabe ksinny jednorozec 16:00:45 Current chairs: dustymabe jednorozec ksinny masta mboddu nirik pbrobinson puiterwijk sharkcz tyll 16:00:45 #topic init process 16:00:59 morning 16:01:08 * sharkcz is here today 16:01:25 * mboddu waves at nirik and sharkcz 16:01:57 So, lets get started 16:02:03 #topic EPEL8 16:03:32 nirik: I am not sure whether I told you the reason why I asked for fxx and fxx-candidate targets. https://pagure.io/fedscm-admin/pull-request/16, look at Stephan comment on epel8-playground-candidate target 16:04:29 uh, looking, but there's a lot of comments there. 16:04:42 So, I talked to Dennis few mins back and he said, they were created for different reasons back in the day and later he merged them back together. 16:05:08 nirik: Sorry, cant give you the link since the PR got rebased later 16:05:19 And the link dies :( 16:05:21 ok 16:05:58 When I say merged I mean both of them serve the same purpose now 16:06:23 right, ok 16:06:46 But fedpkg uses fxx-candidate target to identify the target if no target is specified while package.cfg has fxx targets listed 16:06:50 but fedpkg is still... right. 16:07:22 so lets just use both now and if we clean this up later we can fix it everywhere (which will be a big pain, but oh well) 16:07:33 So, should we create epel8-playground target and keep continuing it, or should we ask people to change and start using fxx-candidate 16:07:54 nirik: Okay, you answered before I asked the question :D 16:08:08 I will create the target and update the PR after the meeting 16:08:14 ok 16:08:15 And that should be it 16:08:31 and we can start branching and building and then composing. ;) 16:08:36 smooge: ^ FYI, we might be ready to start branching in couple of hours 16:09:00 nirik: Right, but we are holding on composing for a while, until we have some builds 16:09:12 I am guessing may be next week 16:09:13 sure 16:09:20 Or later, up to smooge 16:09:31 we need a number of basethings 16:09:35 Yup 16:10:08 That reminds me, I dont know if you have seen my message couple of days back 16:10:13 * mboddu digs up the logs 16:11:09 [09:22:34] nirik: You remember the buildroots for epel8 and epel8-playground, I think I found something, not sure if its an issue or something we might want to have 16:11:09 [09:22:58] epel8 buildroot = rhel + epel (stable+overrrides) 16:11:09 [09:23:59] epel8-playground = rhel + epel8-playground + epel8(only stable) 16:11:09 [09:24:57] From what I think it might be desired since when they move stuff epel8-playground to epel8, this might be helpful for them 16:11:34 .here 16:12:08 yeah, we will want that indeed. 16:12:40 i saw I needed to add some epel8 points to koji-gc and was wondering when that should go in. it needs keys 16:12:40 Awesome 16:12:46 oh sorry 16:13:45 smooge: Yeah, we need update koji configs and some scripts 16:15:26 nirik: We need to build fedpkg and fedscm-admin, I kinda own fedscm-admin and I can build it, but fedpkg we might have to add the patch and build it 16:15:52 ok. or can you ask upstream to do a new release? that might be easiest. 16:16:33 nirik: Actually let me send them an email right away, before they go out for the day 16:16:47 * mboddu writes an email, bbiab 16:20:14 Sent the email 16:20:50 Anything else on EPEL8? 16:21:58 nope 16:22:05 Okay, moving on 16:22:15 #topic Alternate Architecture Updates 16:22:26 sharkcz: Any updates? 16:22:46 nirik: I noticed that switching s390x to KVM didn't help us much, it took the same amount of time to run the compose 16:22:52 nope, but thanks to nirik for the new s390x builders, they are very fast even for the images 16:22:54 I didn't look at the image times 16:22:59 yes, I was disappointed as well... 16:23:08 something is taking longer, I hope to look into it. 16:23:23 I think it might be ppc64le 16:23:25 nirik: Yeah, me to :) 16:23:32 yes, it;'s ppc64le 16:23:48 s390x container was ~15 minutes 16:23:53 I was gonna try and update/reboot them again...it might just be storage tho 16:24:11 I guess I could only enable 1/2 of them and see if that helps 16:24:30 enable what? 16:25:09 the builders? there are 9 of them on each power9 box... I could only enable 4 or 5 and see if they do things better without contending on the storage. 16:25:22 nirik: Ahh, got it 16:25:43 or perhaps the problem is something else. I am not sure. 16:26:00 nirik: is it p8 or p9 hw? 16:26:12 it's the p9's that are slow. 16:26:27 yes, then it could be the sata storage 16:26:52 but it could be upgraded IMO 16:27:15 yeah. I asked about that... we need more proof that that is the problem 16:27:51 ack 16:28:20 but I can try just 4 enabled per and do a compose and see? 16:28:43 worth trying, for sure 16:28:48 nirik: ack 16:29:02 are they used for the images? so with nested virt? 16:29:03 but thats only 3 of them in the image channel. 16:29:18 yeah, nested virt. 16:30:08 so maybe I/O inside nested virt ... 16:30:25 it's getting complex 16:30:28 https://koji.fedoraproject.org/koji/taskinfo?taskID=36293914 16:30:33 In recent container base image - https://koji.fedoraproject.org/koji/taskinfo?taskID=36297774 16:30:36 2+ hours for a ostree? 16:30:50 s390x completed in 5 min, ppc64le took almost and hour and failed 16:31:49 s/and/an/ 16:32:00 so yeah, wonder if the nested virt is not happy somehow 16:32:23 host has options kvm_hv nested=1 16:32:36 the cpu will be proabbly OK, but I/O inside the nested VM can be the problem 16:32:53 guests have: 16:32:56 16:32:56 16:32:56 16:33:24 so... hum... 16:33:28 what's the storage setup? LVs or image files? 16:33:48 * mboddu wishes oz throws some timestamps in the logs 16:34:07 oh, right I can't do that. I was thinking I could just go back to the p8 ones... but we need the p9 ones or qemu has issues. 16:34:13 sharkcz: lv's 16:34:26 hmm 16:34:43 it's got so much space... sadly it's 7200rpm sata. 16:34:51 /dev/md2 vg_guests lvm2 a-- <21.83t 20.54t 16:35:00 buildvm-ppc64le-01.ppc.fedoraproject.org vg_guests -wi-ao---- 141.60g 16:35:04 but I guess we can try to replicate the process outside koji on another p9 and will see 16:35:32 yeah, that would be great if can... 16:36:17 we have a boston p9 in brno, will plan for some testing there 16:36:25 sharkcz: Do you have a p9 box? 16:36:29 or I can test at home :-) 16:36:45 * mboddu is jealous of sharkcz :P 16:36:47 so, I am going to update/reboot them, and then let me put some on the second one... 16:36:51 nice 16:36:56 but my disks are even worse, 500G SATAs ;-) 16:36:58 ack 16:37:09 worth going to 5.2 kernel? 16:37:21 works here 16:37:48 but don't think it brings any perf changes 16:39:17 ok 16:39:26 sharkcz: We are using 5.1.16-200.fc29.ppc64le 16:42:12 hum, so I have 01/02 in compose and image. Let me move image to different dedicated ones. Perhaps it's doing runroot tasks at the same time as images on those. 16:43:52 #info sharkcz will be doing some experiments to gather some facts on ppc64le slowness and nirik will 4 builders on each p9 box and see if it helps with any storage competition issue. 16:43:54 so, I'll update/reboot/shuffle channels around and fire a new compose and we can see from there. ;) 16:44:07 nirik: ack, let me know if you need any help 16:44:23 Moving 01 and 02 out of compose channel? 16:44:30 Or the other way around? 16:44:47 removing 01 and 02 from image and add couple more there? 16:45:01 the second. 16:45:09 Awesome 16:45:12 so, 01/02 only in compose, 03/04/10/11 in image 16:45:22 Oh 4 for image, nice 16:47:09 #info nirik will remove buildvm-ppc64le-01/02.ppc.fedoraproject.org ppc64le builders from image channel and add 04,10,11 ppc64le builders to image channel which makes 4 in image channel including 03 16:51:15 wow... checkout the screenshot here: https://koji.fedoraproject.org/koji/taskinfo?taskID=36284507 16:52:27 that doesn't look right ... 16:53:24 anyhow... 16:54:15 Yeah, I wonder whats causing it 16:54:38 Anyway, anything else? 16:54:49 sharkcz, nirik the recommendation I got from an ibm'ers was that the p9 on sata needed to use hardware raid or it bypassed the onboard cache in jbod mode. 16:55:41 that hardware raid 'should' be raid6 but even raid0 with 1 drive per raid0 should enable the cache 16:56:08 smooge: I find that pretty hard to beleive, but ok 16:56:19 i don't know how valid that advise is but it is a controller item 16:56:22 linux claims the controller is using cache 16:57:36 Write cache: enabled, read cache: enabled, supports DPO and FUA 16:58:02 on the other hand if JBOD is just a pass-thru, so it won't cache on the controller 16:58:19 nirik: isn't it just the disk's cache? 16:58:35 could be. 16:58:46 it may be a multiple definition of using the cache. Again.. that is what I got told as a probable reason for the poor performance 16:58:52 I would say for higher performance you will need cache from the I/O controller too 16:59:04 whats the disk benchmark to use these days? 16:59:32 no idea :-) 16:59:37 maybe hdparm still 17:01:43 the "use the raid function" makes sense to me 17:02:07 to some degree 17:02:59 Okay, we passed the scheduled time, lets take the remaining discussion to #fedora-releng 17:03:03 Thanks guys for joining 17:03:11 #endmeeting