Okay, more git bashing...
After losing 3 hours' work today from something as simple as "git stash
save" (where git stashed 3000+ untracked/generated files, despite the docs
saying it doesn't do that), then not being able to do "stash apply"
(because "file already exists..." for 3000+ files), and having to construct
the work from screenshots of the diff i was able to grab before deleting
and checking out again (as one so often has to do with git), i went
searching for "data loss in git" and stumbled across this page near the top
of the results:
What really makes that worth reading is the list of suggestions at the end
of the page. They start out with this little gem:
Internalize the concept that git is *designed* to forget things. If you
haven't seen something reach another repository, maybe it didn't. Heck,
even if you *did* see it go somewhere else, maybe it fell out of the
historical record there and then got garbage-collected.
It makes me sick to no end that people accept that so readily, and then go
back for a second helping.
It occurred to me today that in nearly 31 years of using a computer i have,
in total, lost more data to git (while following the instructions!!!) than
any other single piece of software. Also concluded is that git is the only
SCM out there which makes SCM difficult for the simple stuff. Even RCS is
simpler to use. Sure CVS has limits, but respect those limits and it works
just fine. Never lost a line of code in CVS.
----- stephan beal
"Freedom is sloppy. But since tyranny's the only guaranteed byproduct of
those who insist on a perfect world, freedom will have to do." -- Bigby Wolf
fossil-users mailing list