20:00:09 #startmeeting 20:00:12 * ricky 20:00:21 #topic Infrastructure -- Who's here? 20:00:35 * ricky 20:00:41 * LinuxCode  20:00:44 * SmootherFrOgZ is 20:00:45 * lmacken 20:00:48 * mmcgrath is here 20:00:50 * nirik is lurking around. 20:01:17 Ok, well lets get started 20:01:24 #topic Infrastructure -- F11 Release 20:01:40 .ticket 1395 20:01:43 mmcgrath: #1395 (Lessons Learned) - Fedora Infrastructure - Trac - https://fedorahosted.org/fedora-infrastructure/ticket/1395 20:01:51 So what went wrong with this release? 20:02:02 All in all I think it went very well. 20:02:04 very little I think. ;) 20:02:20 The early bit flip (thanks f13) provided a 75% mirror rate when the anouncement went out. 20:02:24 * herlo is lurking 20:02:27 thats much better then for F10. 20:02:29 I didn't expect docs links to change, so we had to do some last-minute updates that made us a bit late 20:02:38 We'll make sure to double check that better next time around. 20:02:48 Also, as mdomsch has pointed out many times, the torrents were screaming. The fastest torrent dl's I've ever had during a release. 20:03:01 * sijis is here. 20:03:03 15-20 minutes for 4 isos compared to 8 or 9 hours previously. 20:03:04 * ianweller 20:03:11 * abadger1999 here 20:03:17 yeah, due to them opening early we got a lot more seeders at release time. 20:03:23 ricky: should we make that a docs task or an us task? 20:03:45 That'll be a websites task, I'll send links to the test page to list in advance. 20:04:29 * johe_ here too, by the way 20:04:29 k 20:04:52 The other thing that went kind of wacky was the bit leaks we had the thursday or friday before hand. 20:04:57 I think f13 is working on those though. 20:05:16 Oh, one other thing. 20:05:18 Proxy3 was hurting. 20:05:26 And I think I know why but I need to do a bit more research. 20:05:30 long story short it was doing too much. 20:05:41 I think it's just because it's one of our slower boxes in terms of cpu and disk. 20:05:49 Did the x86_64 vs. i386 difference also come into play? 20:06:06 ricky: not this time. But proxy1 and proxy2 are kind of odd because they're behind a balancer. 20:06:07 * mmcgrath needs to fix that. 20:06:14 Ah 20:06:37 But once the release was out the door.... Things went great. 20:06:42 Anyone have anything else on this topic? 20:06:42 why are they behind the balancer, but the slowest box has to face the cold cruel world all alone? 20:07:04 rjune_wrk: well, we've got 4 sites, only one of which has a balancer. 20:07:19 so we've got 5 proxy servers, 2 behind the balancer, the others are all at the various sites. 20:07:23 ah, ok. 20:07:44 Ok, anyone have anything else? 20:08:00 Solid 20:08:08 #topic Infrastructure -- Fedora Community 20:08:12 We deployed our first TurboGears2 application this week :) 20:08:17 So fedora community shipped. 20:08:22 lmacken: you want to talk a bit about all of it? 20:08:29 Sure 20:08:36 \o/ 20:08:37 Software speaking, not much went wrong. 20:08:37 * mmcgrath is talking about https://admin.fedoraproject.org/community/ 20:08:40 I hit some insane traceback during staging that was caused by pointing to the unresolvabl 20:08:45 e db1.stg, instead of just db1. Other than that, everything went very smooth. 20:09:03 It's using both of our new memecached servers, using postgres on db2 for the feed store, and currently load balancing to mod_wsgi on app1 and 2 20:09:22 and it's the snappiest I've ever seen the app as of yet 20:09:39 but anyway, I did another update today to fix a variety of bugs, and we're going to keep cranking on it 20:09:42 It also means we can, at our leasure, convert our apps to TG2. 20:09:50 I want to plan a post-mortem when spot gets back from vacation 20:10:02 so we can talk about what happened, and how we can do it better 20:10:20 lmacken: is the whole tg2 stack in Fedora yet? I need to install and start playing with it. 20:10:39 the whole TurboGears2/FedoraCommunity/Moksha stack is in our fedora-infrastructure repo, and a chunk of it is still under review 20:10:50 Does our tg2 support mean close-to automatic support for pylons as well? 20:10:57 ricky: yes 20:11:00 Awesome 20:11:03 pylons should work fine now 20:11:12 excellent. 20:11:15 Anyone have any questions for Luke on fedora community? 20:11:20 I will start writing up a TurboGears2 SOP 20:11:45 I'd like to get together some questions to ask spot wrt licensing. 20:12:14 I guess that can be discussed at the post-mortem as soon as spot is back - can that be on IRC? 20:12:15 That way we can have less round trips to legal. 20:12:24 IRC +1 20:12:41 I just want to be better educated about it before I copy and paste :) 20:12:49 yes, it will be public, probably on irc or asterisk 20:13:20 IRC+1 20:13:28 Anyone have anything else on this? 20:13:45 asterisk -1; IRC is better for public stuff. 20:13:47 mmcgrath: when do you wanna spread to app3-6? 20:13:50 lmacken: good work, thanks for the time you've put in on it. 20:13:51 abadger1999: agreed 20:14:00 lmacken: today or tomorrow. I know you did some updates, were they straight rpm updates? 20:14:22 mmcgrath: thank you. Thanks for your help as well, couldn't have done it without you guys 20:14:51 mmcgrath: yep, straight rpm updates 20:14:55 coolz. 20:14:58 ok so next topic 20:15:05 #topic Infrastructure -- Memcached 20:15:16 we've been slipping down the slippery slope that is memcached. 20:15:21 a month ago we were using it for sessions on the wiki. 20:15:41 starting today we've got 3 apps using it and dedicated memcached servers. 20:16:02 the fas memcached stuff is going to exist just on the fas servers for security reasons. 20:16:09 everything else is on memcached[1-2] 20:16:21 ricky: do you want to remove memcached from puppet on app[1-2] or do you want me to 20:16:24 ? 20:16:29 I'll get that after the meeting 20:16:38 solid 20:16:47 Nothing major here, just mentioning it. Any questions? 20:16:56 Speaking of security, I just want to let everybody know that load access into memcached = arbitrary command execution with python-memcached 20:17:07 So be careful about that :-) 20:17:28 20:17:52 k 20:18:01 #topic Infrastructure -- fas update 20:18:12 So we just got done (well, are still doing) the fas updates. 20:18:14 This was a pretty major change as changes go. 20:18:23 Especially in how our shell accounts and aliases are generated. 20:18:52 So keep your eyes out for oddities. 20:19:10 So far things have been going fine, we've been running in staging for a while now without issue. 20:19:30 I'm going to be moving our fas-sync to run more often on some of our more public servers 20:19:35 fedorapeople, fedorahosted, cvs, etc. 20:19:43 Any questions there? 20:19:51 Woohoo :-) My goal is <= 5 minutes sync time 20:20:11 :) 20:20:12 Ok 20:20:19 #topic Infrastructure -- Outage this morning 20:20:26 So we had some doom this morning for those of you hanging out in #sysadmin-noc 20:20:35 Not sure what happened yet but I opened a bug about it 20:20:36 https://bugzilla.redhat.com/505352 20:20:42 xen13 up and died. 20:20:46 and took db2 with it, twice. 20:20:53 I moved db2 to a different host, hasn't happened since. 20:21:10 at the moment, nothing new on this so I'll move on. 20:21:19 #topic Infrastructure -- Pkgdb update 20:21:27 abadger1999: we're on for a pkgdb rollout on Monday? 20:21:32 that's right. 20:21:47 and this is a wsgi conversion? 20:21:50 any other changes? 20:21:57 It's a major update that gets us using wsgi for the packagedb, all the bugfixes that accrued during change freeze, 20:22:11 I've still got smolt to convert. 20:22:11 moving to usernames instead of userids i nthe public API. 20:22:18 after that I *think* all of our tg stuff will be good to go. 20:22:24 Afdter pkgdb and smolt, will all of our apps be on haproxy + wsgi? 20:22:32 Switching package status from the orphan owner to the status field. 20:22:38 Lots of stuff changing. 20:23:35 I'll get a new bodhi ready to go at that time 20:23:44 I should probably get a new one into staging now so we can test bodhi#319 20:23:45 ricky: I believe so. 20:24:00 lmacken: Excellent 20:24:03 lmacken: sounds good. 20:24:06 Anything else on this? 20:24:23 Test out admin.stg.fedoraproject.org/pkgdb ! ;-) 20:24:31 hehehe 20:24:38 #topic Infrastructure -- releng1.stg 20:24:40 I'm sure there's more bugs since this has major changes. 20:24:43 After the pkgdb update, let's go through staging 20:24:44 speaking of staging 20:24:47 abadger1999: yeah :) 20:25:00 releng1.stg is nearing rediness. 20:25:01 And make sure to get all the commits we want out of it before restarting the branch. 20:25:25 lmacken: I need to find out from you and jwb what is missing from it. I think it's going to take a close eye to make sure we know what we're putting in staging. 20:25:32 cron jobs and things that might fail, that sort of thing. 20:25:47 mmcgrath: readonly mount of /mnt/koji? 20:25:52 lmacken: it'll have that. 20:26:10 ok, to test mashing and whatnot, we'll have to tweak some paths 20:26:35 lmacken: and I don't think anything mash like is installed yet. 20:26:47 but I figure we'll work through it one boy one until it's usable for you and jwb. 20:27:00 ok, cool 20:27:01 I also assume you're going to need some sort of $TEMP_STORAGE 20:27:07 yeah 20:27:14 that's fairly large. we'll figure that out when the time comes. 20:27:19 was jwb at FAD? 20:27:24 yep 20:27:33 probably traveling / busy 20:27:36 we'll get ahold of him. 20:27:38 ok, ricky :) 20:27:44 #topic Infrastructure -- Staging 20:28:00 So we need to refresh the staging branch at some point soon. 20:28:25 and make sure all of our staging commits get merged back with master. 20:28:37 ricky: what are a rundown of some of the commits left? 20:28:40 I know pkgdb had a bunch. 20:29:07 Here's the full diff: http://ricky.fedorapeople.org/staging.diff 20:29:25 It's mostly pkgdb, I thnk 20:29:27 **think 20:29:55 It's not that bad, so we might get away with not blowing it away 20:30:07 But I guess that's something to look at once pkgdb changes have been merged. 20:30:11 pkgdb will be synced after Monday, so that's good. 20:30:14 Yeah 20:30:20 Cool. 20:30:25 Do we have a good recipe for merging the changes yet? 20:30:32 And thanks to mmcgrath for starting to merge often this week 20:30:37 All in all, considering we don't really monitor this interaction much... things are in better shape then they otherwise could be. 20:30:56 ricky: figured i'd make a habit of it 20:31:42 abadger1999: It'll probably consist mostly of finding out what we want to keep or throw away 20:32:08 The process of merging isn't that painful at all, just getting updates on half-finished or abandoned work in staging. 20:32:30 ricky: I'm just wondering, can I copy changes from staging to production, commit, and git will figure it out? Or do I need to use some git command to pick out specific changesets? 20:33:17 It's cleaner to cherry-pick, but I *think* hand-copying will work as well. I haven't tried doing work that way though, so I'm not sure how well it gets handled. 20:33:27 I think the diff is short enough for us to go through it manually though. 20:33:39 I've hand-copied with success in the past. 20:33:47 but try to use git just so I'm mo'betta with it. 20:35:15 Okay. We'll see what it looks like after Monday's merge ;-) 20:35:15 Ok, anything else on that? 20:35:18 yeah 20:35:31 With that, we'll open the floor 20:35:37 #topic Infrastructure -- Open Floor 20:35:43 Anyone have anything they'd like to discuss? 20:35:46 apart from a million questions I have ? 20:35:49 lol 20:35:59 Ask away :-) 20:36:01 I will have to figure out what is done for what purpose the enxt weeks 20:36:20 ricky, I ask you guys the next weeks, or we would be here all night 20:36:25 lol 20:36:32 feel free to ask in #fedora-admin after the meeting as well 20:36:38 Id like to figure out how everything fits together 20:36:41 sure 20:36:59 I have one question though, the fas update was to kill bugs ? 20:37:08 it said so in the email I had 20:37:18 LinuxCode: For some definition of bugs, yes. 20:37:40 abadger1999, I take it to make sure that people participate and not just sign up then ? 20:37:41 Ther were bugs and there was also performance problems that were causing failures to occur. 20:37:46 ohh ok 20:37:50 i see 20:38:17 The extra @fp.o aliases being created was a bug too 20:38:26 ricky, thats what I meant 20:38:27 Unfortunately, it went unnoticed for a loong time. 20:38:47 would have been nice to give people a week though to sort themselves out 20:38:56 The policy hasn't changed, it was just a bug that we weren't enforcing it. 20:38:59 I had the email today saying about the bug 20:39:04 hmm k 20:39:13 well it doesnt matter now 20:39:21 for me anyway 20:39:23 ;-} 20:39:25 heheh 20:39:33 Anyone have anything else to discuss? If not we'll close the meeting early. 20:40:04 Closing in 30 20:40:34 10 20:40:48 #endmeeting