On Wed, 21 May 2014 00:48:48 -0700 Dan Williams wrote: > On Fri, May 16, 2014 at 8:04 AM, Chris Mason wrote: > > -----BEGIN PGP SIGNED MESSAGE----- > > Hash: SHA1 > > > > On 05/15/2014 10:56 PM, NeilBrown wrote: > >> On Thu, 15 May 2014 16:13:58 -0700 Dan Williams > >> wrote: > >> > >>> What would it take and would we even consider moving 2x faster > >>> than we are now? > >> > >> Hi Dan, you seem to be suggesting that there is some limit other > >> than "competent engineering time" which is slowing Linux "progress" > >> down. > >> > >> Are you really suggesting that? What might these other limits be? > >> > >> Certainly there are limits to minimum gap between conceptualisation > >> and release (at least one release cycle), but is there really a > >> limit to the parallelism that can be achieved? > > > > I haven't compared the FB commit rates with the kernel, but I'll > > pretend Dan's basic thesis is right and talk about which parts of the > > facebook model may move faster than the kernel. > > > > The facebook is pretty similar to the way the kernel works. The merge > > window lasts a few days and the major releases are every week, but > > overall it isn't too far away. > > > > The biggest difference is that we have a centralized tool for > > reviewing the patches, and once it has been reviewed by a specific > > number of people, you push it in. > > > > The patch submission tool runs the patch through lint and various > > static analysis to make sure it follows proper coding style and > > doesn't include patterns of known bugs. This cuts down on the review > > work because the silly coding style mistakes are gone before it gets > > to the tool. > > > > When you put in a patch, you have to put in reviewers, and they get a > > little notification that your patch needs review. Once the reviewers > > are happy, you push the patch in. > > > > The biggest difference: there are no maintainers. If I want to go > > change the calendar tool to fix a bug, I patch it, get someone else to > > sign off and push. > > > > All of which is my way of saying the maintainers (me included) are the > > biggest bottleneck. There are a lot of reasons I think the maintainer > > model fits the kernel better, but at least for btrfs I'm trying to > > speed up the patch review process and use patchwork more effectively. > > To be clear, I'm not arguing for a maintainer-less model. We don't > have the tooling or operational-data to support that. We need > maintainers to say "no". But, what I think we can do is give > maintainers more varied ways to say it. The goal, de-escalate the > merge event as a declaration that the code quality/architecture > conversation is over. > > Release early, release often, and with care merge often. I think this falls foul of the "no regressions" rule. The kernel policy is that once the functionality gets to users, it cannot be taken away. Individual drivers in 'staging' manage to avoid this rule because that are clearly separate things. New system calls and attributes in sysfs etc seem to be much harder to "partially" release. To quote from Linus in a recent interview: "I personally think the stable development model is not one of continual incremental improvements, but a succession of overshooting and crashing." Yet Linux is stuck in "incremental improvement" mode and is not in a position to "overshoot and crash" much. I agree there is a problem. I can't see a solution. NeilBrown