Warly wrote:
>Mandrake Linux is based on a fixed date releasing due to market
>constraints (distributor/supplier schedules). As a consequence the
>release date was scheduled 3 months ago. On this date we have a
>very little margin of 2 or 3 days, but not more.
>
>So the datum is: "release date is March the 15th".
>
>Now we need to do it in this timeframe (and we are the 17th), and
>we have no other choice.
>
>In a few days 8.2 updates will be released, and them you will have
>the real stable and polished distro you want.
>
>Debian has no real realease date, as a consequence doing a stable
>release is just a piece of cake, anybody can do it.
>
>Maybe our model is not good.
>
>Thinking of a better model, I am not sure what I could choose. No
>deadline means no hurry, and "if the last moments didn't exist,
>nothing good would be done".
>
>Moreover if we wait too much, new versions of nearly all the tools in
>the distro will be released, and their respective authors will not
>care anymore fixing bugs in the old version we would have included.
>
>I do think that our model is not so bad, and that we are reaching a
>good compromise between cutting edge and stability, but /this/
>compromise _is needed_.
>
There is an alternative model, but it requires a high initial investment.
Bugs in packaging are usually monor but some of those can cause
significant problems in Mandrake tools... An improper script to set up
dotfiles for defaults in the security script, or a change in the rpm
tool and how it treats the formation of such files under update, scould
render the Mandtrake Upgrade tool useless from distro to distro. In
this area, a bot that recognizes such scripts in the packagoing, and
someone that examines the results as a duty could help a lot.
Testing, per se, is not a solution to better quality. Microsoft throws
much money at it, shadows every programmer with a designated tester.
They even have scholarly types who move pins on maps from alpha to beta
to gamma and pigeonhole all the test categories. Their results should
speak for themselves.
The real effort needs to go into planning tools and programs, not just
the generalities, but the details. This needs to be done by developers
who are trained to interact constructively. The training alone is a
killer. I am talking about a bill of 3 weeks with no production, just
training.
And that is only half the bill. The other half is putting the training
to use. It would mean that bug-barriers and unanticipated cases count
would drop dramatically.
As other posters suggest, that requires communication, and that means
internal communication, among developers, using the best conference
tools available, since even Mandrake developers are scattered a bit.
Obviously, any such model will have to wait until Mandrake turns the
corner on profitability, but this model is far less expensive that the
Microsoft one or the lockstep of "alpha", "beta" releases.
Basically the release schema would look like this:
Developers Training-->Input Objectives for Next release->Developers
confer to design tools to meet objectibves->
-->First programming on _new_ tools with no GUI work
\
\
\-> Testers input on hardware and software unanticipated cases for tools
------> Developers work on other projects while reports collect,
probably other tools or
GUI-ing tools that seem to work well
\
\
\->Testers ask for other packages to test, seeming to have found
most of the cases
------>Developers work on fixes and release the secondary projects they
were working on
\
\
\->Continue the case mismatch stuff
----->Release second round of non-GUI scratch editions of tools
\
\
\->Offer brainstorming ideas about making it user-friendly and
setting up the GUI
(no arguing--no bad ideas-go for quantity of ideas collect and
group them later)
(do not discuss ideas or comment on them--building on them is OK)
-----> Conference by best tools available with testers and sort out the
current best ideas
for the design.
\
\
\->Finish testing second round, review the product
specifications/goals from the
Brainstorming/multivote process to get a good idea what to test
for on the GUI
----->Finish on minor bugs on the non-GUI versions of the tools--Go for
first-edition GUI
\
\
\->Test with brainlessness and abandon... Do everything wrong that
can be imagined for
a total newbie or a windows convert. (Even include installing
to hdc and then
removing it and moving hdc to hda and see how the software
handles it)
Don't laugh. There is an actual support case where that happened.
----->Release some tools and continue work on others. Continue
conferences on each
tool and make sure that internally the knowledge of all is
shared in making these
tools.
Now I know there is some loss in time and effort with coding the action
separately from the interface or in fact coding a text as well as a
graphic interface, but consider that both can be offered as features....
And the other time loss is the training. The conferencing is part of the
cost of the product and should not be considered a loss. The expulsion
of bugs is te payoff, as well as better-designed software. If we had
thought up front about broken connections and tested for them while
developing Software manager, it would be at its current state 2
generations ago. That would have been worth a lot more than the
conferencing cost/time loss.
The barrier to this model is financial. mandrakesoft cannot easily
afford to blow three weeks of every developer's time for the training
nor easily work through the initial roughnesses of the new system. It
would require steadfast leadership and extreme effort on the part of
management to make it work. When the finances are better you might try
to implement it, but for now, with the exception of better communication
to testers, I think you are doing the best that can be done.
Civileme