cloudstack-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Rajani Karuturi <>
Subject Re: Release cadence
Date Wed, 19 Mar 2014 11:45:46 GMT
The primary problem I feel is that we dont plan our releases.(I am fairly new here and I may
be wrong)
The role of the release manager starts only during the RC creation phase asking for votes(again
I maybe wrong).

I feel it should start much earlier. Everyone who is actively involved should have a clear
idea on what we are going to release next.
We do this to an extent by sending proposals etc. but we really dont do planning and the big
picture is missing. 

I think the RM should facilitate planning of the release and in the process take feedback
from the committers and users who are going to work on the release. 
A single view of the current release status also would help immensely. 

I like the way does the release management through trello boards[1]. Probably
we could explore such options. It facilities voting on features, gives a quick view of whats
getting in the release which would help QA/anyone interested in testing to plan the testing
early on instead of waiting for RC.

I am for for short release cycles. "Release early, release often"
It shows project activity and builds developer confidence as long as we do quality releases



On 17-Mar-2014, at 10:22 pm, John Kinsella <> wrote:

> I am in agreement with my radical CloudStack brother.
> On Mar 13, 2014, at 9:42 AM, David Nalley <> wrote:
>> The RC7 vote thread contained a lot of discussion around release
>> cadence, and I figured I'd move that to a thread that has a better
>> subject so there is better visibility to list participants who don't
>> read every thread.
>> When I look at things schedule wise, I see our aims and our reality.
>> We have a relatively short development window (in the schedule) and we
>> have almost 50% of our time in the schedule allocated to testing.
>> (over two months). However, it seems that a lot of testing - or at
>> least a lot of testing for  what became blockers to the release didn't
>> appear to happen until RCs were kicked out - and that's where our
>> schedule has fallen apart for multiple releases. The automated tests
>> we have were clean when we issued RCs, so we clearly don't have the
>> depth needed from an automated standpoint.
>> Two problems, one cultural and one technical. The technical problem is
>> that our automated test suite isn't deep enough to give us a high
>> level of confidence that we should release. The cultural problem is
>> that many of us wait until the release period of the schedule to test.
>> What does that have to do with release cadence? Well inherently not
>> much; but let me describe my concerns. As a project; the schedule is
>> meaningless if we don't follow it; and effectively the release date is
>> held hostage. Personally, I do want as few bugs as possible, but it's
>> a balancing act where people doubt our ability if we aren't able to
>> ship. I don't think it matters if we move to 6 month cycles, if this
>> behavior continues, we'd miss the 6 month date as well and push to 8
>> or 9 months. See my radical proposition at the bottom for an idea on
>> dealing with this.
>> I also find myself agreeing with Daan on the additional complexity.
>> Increasing the window for release inherently increases the window for
>> feature development. As soon as we branch a release, master is open
>> for feature development again. This means a potential for greater
>> change at each release. Change is a risk to quality; or at least an
>> unknown that we again have to test. The greater that quantity of
>> change, the greater the potential threat to quality.
>> Radical proposition:
>> Because we have two problems, of different nature, we are in a
>> difficult situation. This is a possible solution, and I'd appreciate
>> you reading and considering it.  Feedback is welcome. I propose that
>> after we enter the RC stage that we not entertain any bugs as blockers
>> that don't have automated test cases associated with them. This means
>> that you are still welcome to do manual testing of your pet feature
>> and the things that are important to you; during the testing window
>> (or anytime really). However, if the automation suite isn't also
>> failing then we consider the release as high enough quality to ship.
>> This isn't something we can codify, but the PMC can certainly adopt
>> this attitude as a group when voting. Which also means that we can
>> deviate from it. If you brought up a blocker for release - we should
>> be immediately looking at how we can write a test for that behavior.
>> This should also mean several other behaviors need to become a valid
>> part of our process. We need to ensure that things are well tested
>> before allowing a merge. This means we need a known state of master,
>> and we need to perform testing that allows us to confirm that a patch
>> does no harm. We also need to insist on implementation of
>> comprehensive tests for every inbound feature.
>> Thoughts, comments, flames, death threats? :)
>> --David

View raw message