@@ -566,6 +566,7 @@ enum qed_slowpath_wq_flag {
QED_SLOWPATH_MFW_TLV_REQ,
QED_SLOWPATH_PERIODIC_DB_REC,
QED_SLOWPATH_ACTIVE,
+ QED_SLOWPATH_RESCHEDULE,
};
struct qed_hwfn {
@@ -1098,10 +1098,13 @@ static int qed_slowpath_delayed_work(struct qed_hwfn *hwfn,
if (!test_bit(QED_SLOWPATH_ACTIVE, &hwfn->slowpath_task_flags))
return -EINVAL;
- /* Memory barrier for setting atomic bit */
- smp_mb__before_atomic();
- set_bit(wq_flag, &hwfn->slowpath_task_flags);
- smp_mb__after_atomic();
+ if (wq_flag != QED_SLOWPATH_RESCHEDULE) {
+ /* Memory barrier for setting atomic bit */
+ smp_mb__before_atomic();
+ set_bit(wq_flag, &hwfn->slowpath_task_flags);
+ smp_mb__after_atomic();
+ }
+
queue_delayed_work(hwfn->slowpath_wq, &hwfn->slowpath_task, delay);
return 0;
@@ -1155,8 +1158,8 @@ static void qed_slowpath_task(struct work_struct *work)
if (!ptt) {
if (test_bit(QED_SLOWPATH_ACTIVE, &hwfn->slowpath_task_flags))
- queue_delayed_work(hwfn->slowpath_wq,
- &hwfn->slowpath_task, 0);
+ qed_slowpath_delayed_work(hwfn,
+ QED_SLOWPATH_RESCHEDULE, 0);
return;
}