Why would you compile the kernel without optimization?
Can another reason be occasionally still relevant?
Will the compilation be a bit quicker when extra data processing could be omitted?
There's many places in the kernel that WILL NOT BUILD without optimization.
Would you like to keep the software situation in this way?
In fact, we do a lot of tricks to make sure that things work the way we expect it to, because we add broken code that only gets compiled out when gcc optimizes the code the way we expect it to be, and the kernel build will break otherwise.
* Can this goal be also achieved without the addition of “broken code”?
* How do you think about to improve the error handling there?
Regards, Markus -- To unsubscribe from this list: send the line "unsubscribe linux-kselftest" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html