On 30/06/2022 17:01, Mauro Carvalho Chehab wrote:
Em Thu, 30 Jun 2022 09:12:41 +0100 Tvrtko Ursulin tvrtko.ursulin@linux.intel.com escreveu:
On 30/06/2022 08:32, Mauro Carvalho Chehab wrote:
Em Wed, 29 Jun 2022 17:02:59 +0100 Tvrtko Ursulin tvrtko.ursulin@linux.intel.com escreveu:
On 29/06/2022 16:30, Mauro Carvalho Chehab wrote:
On Tue, 28 Jun 2022 16:49:23 +0100 Tvrtko Ursulin tvrtko.ursulin@linux.intel.com wrote:
.. which for me means a different patch 1, followed by patch 6 (moved to be patch 2) would be ideal stable material.
Then we have the current patch 2 which is open/unknown (to me at least).
And the rest seem like optimisations which shouldn't be tagged as fixes.
Apart from patch 5 which should be cc: stable, but no fixes as agreed.
Could you please double check if what I am suggesting here is feasible to implement and if it is just send those minimal patches out alone?
Tested and porting just those 3 patches are enough to fix the Broadwell bug.
So, I submitted a v2 of this series with just those. They all need to be backported to stable.
I would really like to give even a smaller fix a try. Something like, although not even compile tested:
commit 4d5e94aef164772f4d85b3b4c1a46eac9a2bd680 Author: Chris Wilson chris.p.wilson@intel.com Date: Wed Jun 29 16:25:24 2022 +0100
drm/i915/gt: Serialize TLB invalidates with GT resets Avoid trying to invalidate the TLB in the middle of performing an engine reset, as this may result in the reset timing out. Currently, the TLB invalidate is only serialised by its own mutex, forgoing the uncore lock, but we can take the uncore->lock as well to serialise the mmio access, thereby serialising with the GDRST. Tested on a NUC5i7RYB, BIOS RYBDWi35.86A.0380.2019.0517.1530 with i915 selftest/hangcheck. Cc: stable@vger.kernel.org Fixes: 7938d61591d3 ("drm/i915: Flush TLBs before releasing backing store") Reported-by: Mauro Carvalho Chehab <mchehab@kernel.org> Tested-by: Mauro Carvalho Chehab <mchehab@kernel.org> Reviewed-by: Mauro Carvalho Chehab <mchehab@kernel.org> Signed-off-by: Chris Wilson <chris.p.wilson@intel.com> Cc: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com> Acked-by: Thomas Hellström <thomas.hellstrom@linux.intel.com> Reviewed-by: Andi Shyti <andi.shyti@intel.com> Signed-off-by: Mauro Carvalho Chehab <mchehab@kernel.org> Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
diff --git a/drivers/gpu/drm/i915/gt/intel_gt.c b/drivers/gpu/drm/i915/gt/intel_gt.c index 8da3314bb6bf..aaadd0b02043 100644 --- a/drivers/gpu/drm/i915/gt/intel_gt.c +++ b/drivers/gpu/drm/i915/gt/intel_gt.c @@ -952,7 +952,23 @@ void intel_gt_invalidate_tlbs(struct intel_gt *gt) mutex_lock(>->tlb_invalidate_lock); intel_uncore_forcewake_get(uncore, FORCEWAKE_ALL);
spin_lock_irq(&uncore->lock); /* serialise invalidate with GT reset */
for_each_engine(engine, gt, id) {
struct reg_and_bit rb;
rb = get_reg_and_bit(engine, regs == gen8_regs, regs, num);
if (!i915_mmio_reg_offset(rb.reg))
continue;
intel_uncore_write_fw(uncore, rb.reg, rb.bit);
}
spin_unlock_irq(&uncore->lock);
for_each_engine(engine, gt, id) {
struct reg_and_bit rb;
/* * HW architecture suggest typical invalidation time at 40us, * with pessimistic cases up to 100us and a recommendation to
@@ -960,13 +976,11 @@ void intel_gt_invalidate_tlbs(struct intel_gt *gt) */ const unsigned int timeout_us = 100; const unsigned int timeout_ms = 4;
rb = get_reg_and_bit(engine, regs == gen8_regs, regs, num); if (!i915_mmio_reg_offset(rb.reg)) continue;struct reg_and_bit rb;
intel_uncore_write_fw(uncore, rb.reg, rb.bit); if (__intel_wait_for_register_fw(uncore, rb.reg, rb.bit, 0, timeout_us, timeout_ms,
This won't work, as it is not serializing TLB cache invalidation with i915 resets. Besides that, this is more or less merging patches 1 and 3,
Could you explain why you think it is not doing exactly that? In both versions end result is TLB flush requests are under the uncore lock and waits are outside it.
Sure, but patch 2/3 (see v2) serializes i915 reset with TLB cache changes. This is needed in order to fix the regression.
Not "the" regression, and not even _a_ *regression*. 2/3 fixes an pre-existing and unrelated problem. Or only tangentially related if you want. 2/3 fixes a hang if two engine resets would happen to coincide. Nothing about TLB flushing.
placing patches with different rationales altogether. Upstream rule is to have one logical change per patch.
I don't think it applies in this case. It is simply splitting into two loops so lock can be held across all mmio writes. I think of it this way
- what is the rationale for sending only the first patch to stable? What
does it _fix_ on it's own?
There's no -stable rule enforcing that only one patch would be allowed, nor saying that patches should be fold, doing multiple changes on as single patch just due to "Fixes" tag.
Well if we want to be pedantic what do stable rules say about adding new features - is skipping idle engines (which is a software concept) a fix or a new optimisation?
So, while several -stable fixes can be done on a single patch, there are fixes that will require multiple patches. That's nothing wrong with that.
Agreed. But the point of my argument is that a) 1st patch does not fix anything on it's own (in relation to the regression), b) is adding improvements which will just be extra work to backport to old kernels.
The only rule is that backports should follow what's merged upstream. So, if, in order to fix a regression, multiple patches are needed upstream, in principle, all of those can be backported if they fit at -stable rules.
As an example, once we backported a patch series on media that had ~20 patches, addressing security issues at the media compat32 logic (media ioctls usually pass structs and some with pointers). As the issue was discovered several years after compat32 got introduced, those 22 patches (some containing compat32 redesigns) had to be backported to all maintained LTS.
In this specific case, fixing the regression requires 3 logical changes:
- Split the loop;
- Add serialize logic to i915 reset;
- use the same i915 reset spinlock to serialize TLB cache invalidation.
Neither one of those logical changes alone would solve the issue. That's why I originally added the same Fixes: to the entire series: basically, any Kernel that has the TLB patch backported will require those three logical changes to be backported too.
That basically will follow what's there at the Kernel process docs:
"If your patch fixes a bug in a specific commit, e.g. you found an issue using ``git bisect``, please use the 'Fixes:' tag with the first 12 characters of the SHA-1 ID, and the one line summary."
Documentation/process/submitting-patches.rst
See, Fixes was originally introduced to be a hint to help stable and distro maintainers to identify how far they need to backport a patch. That's mainly why I placed fixes to the entire series. Yet, the same will also happen, in practice, if we place:
Cc: stable@vger.kernel.org # Up to version 4.4
Greg, Sasha and others -stable/distro maintainers will also have a (much less precise) hint about how far the backport is needed.
If this works it would be least painful to backport. The other improvements can then be devoid of the fixes tag.
From backport PoV, it wouldn't make any difference applying one patch or two. See, intel_gt_invalidate_tlbs() function doesn't exist before changeset 7938d61591d3 ("drm/i915: Flush TLBs before releasing backing store"), so, it shouldn't have merge conflicts while backporting it, maybe except if some functions it calls (or parameters) have changed. On such case, the backport fix should be trivial, and the end result of backporting one folded patch or two would be the same.
Yes a lot of things changed. Not least engine and GT pm code. Note that TLB flushing was backported all the way to 4.4 so any hunk you don't strictly need can and will bite you. I have attached a tarball of patches for you to explore. :) Regards,
Thanks! That's very helpful to check the amount of work. It makes easy to use interdiff and (k)diff3 to check what changed.
From it, the differences between 5.4 and 5.16 at intel_gt_invalidate_tlbs() are really trivial.
On 4.14, the function was added on a different file (intel_gem), and there were a few more API differences, as only gen8 code is there, but again, the changes are trivial: mostly macros/functions were renamed and some function parameters changed.
From 4.9 to 4.14 there were also some changes but they also look trivial.
Kernel 4.4 has some other differences - the loop logic is different, and there's a ring initialization function, but, as version 4.4 is not listed anymore as LTS at kernel.org, we probably need to backport only up to 4.9.
All the above should be affecting patch v2 1/3. Patches v2 2/3 and 3/3 just have spin lock/unlock for the gt uncore spinlock. Those will very likely require some work on Kernels 4.x, but folding (or not) the patches won't really help.
What about intel_engine_pm_is_awake, what will you do with that one?
Regards,
Tvrtko