On Fri, 15 Apr 2005, Linus Torvalds wrote:
> I think I've explained my name tracking worries. When it comes to "how to
> merge", there's three issues:
> - we do commonly have merge clashes where both trees have applied the
> exact same patch. That should merge perfectly well using the 3-way
> merge from a common parent that Junio has, but not your current "bring
> patches forward" kind of strategy.
I think 3-way merge is probably the best starting point, but I think that
there might be value in being able to identify the commits of each side
involved in a conflict. I think this would help with cases where both
sides pick up an identical patch, and then each side makes a further
change to a different part of the changed region (you find out that the
other guy's change was supposed to follow the patch, and don't conflict
> - I _do_ actually sometimes merge with dirty state in my working
> directory, which is why I want the merge to take place in a separate
> (and temporary) directory, which allows for a failed merge without
> having any major cleanup. If the merge fails, it's not a big deal, and
> I can just blow the merge directory away without losing the work I had
> in my "real" working directory.
Is there some reason you don't commit before merging? All of the current
merge theory seems to want to merge two commits, using the information git
keeps about them. It should be cheap to get a new clean working directory
to merge in, too, particularly if we add a cache of hardlinkable expanded
> - reliability. I care much less for "clever" than I care for "guaranteed
> to never do the wrong thing". If I have to fix up some stuff by hand,
> I'll happily do so. But if I can't trust the merge and have to _check_
> things by hand afterwards, that will make me leery of the merges, and
> _that_ is bad.
> The third point is why I'm going to the ultra-conservative "three-way
> merge from the common parent". It's not fancy, but it's something I feel
> comfortable with as a merge strategy. For example, arch (and in particular
> darcs) seems to want to try to be "clever" about the merges, and I'd
> always live in fear.
How much do you care about the situation where there is no best common
ancestor (which can happen if you're merging two main lines, each of which
has merged with both of a pair of minor trees)? I think that arch is even
more conservative, in that it doesn't look for a common ancestor, and
reports conflicts whenever changes overlap at all. Of course, reliability
by virtue of never working without help is not a big win over living in
fear; you always have to check over it, not because you're afraid, but
because it needs you to.
> And, finally, there's obviously performance. I _think_ a normal merge with
> nary a conflict and just a few tens of files changed should be possible in
> a second. I realize that sounds crazy to some people, but I think it's
> entirely doable. Half of that is writing the new tree out (that is a
> relative costly op due to the compression). The other half is the "work".
I think that the time spent on I/O will be overwhelmed by the time spent
issuing the command at that rate. It might matter if you start getting
into merging lots of things at once, but that's more like a minute for a
merge group with 600 changes rather than a second per merge; we could
potentially save a lot of time based of having a bunch of information left
over from the previous merge when starting merge number 2. So 15 seconds
plus half a second per merge might be better than a second per merge in
the case that matters.
*This .sig left intentionally blank*
To unsubscribe from this list: send the line "unsubscribe git" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at http://vger.kernel.org/majordomo-info.html