@@ -200,8 +200,8 @@ static inline unsigned int cpumask_local_spread(unsigned int i, int node)
200
200
for ((cpu) = 0; (cpu) < 1; (cpu)++, (void)mask)
201
201
#define for_each_cpu_wrap (cpu , mask , start ) \
202
202
for ((cpu) = 0; (cpu) < 1; (cpu)++, (void)mask, (void)(start))
203
- #define for_each_cpu_and (cpu , mask , and ) \
204
- for ((cpu) = 0; (cpu) < 1; (cpu)++, (void)mask , (void)and )
203
+ #define for_each_cpu_and (cpu , mask1 , mask2 ) \
204
+ for ((cpu) = 0; (cpu) < 1; (cpu)++, (void)mask1 , (void)mask2 )
205
205
#else
206
206
/**
207
207
* cpumask_first - get the first cpu in a cpumask
@@ -290,20 +290,20 @@ extern int cpumask_next_wrap(int n, const struct cpumask *mask, int start, bool
290
290
/**
291
291
* for_each_cpu_and - iterate over every cpu in both masks
292
292
* @cpu: the (optionally unsigned) integer iterator
293
- * @mask : the first cpumask pointer
294
- * @and : the second cpumask pointer
293
+ * @mask1 : the first cpumask pointer
294
+ * @mask2 : the second cpumask pointer
295
295
*
296
296
* This saves a temporary CPU mask in many places. It is equivalent to:
297
297
* struct cpumask tmp;
298
- * cpumask_and(&tmp, &mask , &and );
298
+ * cpumask_and(&tmp, &mask1 , &mask2 );
299
299
* for_each_cpu(cpu, &tmp)
300
300
* ...
301
301
*
302
302
* After the loop, cpu is >= nr_cpu_ids.
303
303
*/
304
- #define for_each_cpu_and (cpu , mask , and ) \
304
+ #define for_each_cpu_and (cpu , mask1 , mask2 ) \
305
305
for ((cpu) = -1; \
306
- (cpu) = cpumask_next_and((cpu), (mask ), (and )), \
306
+ (cpu) = cpumask_next_and((cpu), (mask1 ), (mask2 )), \
307
307
(cpu) < nr_cpu_ids;)
308
308
#endif /* SMP */
309
309
0 commit comments