@@ -1272,15 +1272,6 @@ arch_atomic_try_cmpxchg(atomic_t *v, int *old, int new)
1272
1272
#endif /* arch_atomic_try_cmpxchg_relaxed */
1273
1273
1274
1274
#ifndef arch_atomic_sub_and_test
1275
- /**
1276
- * arch_atomic_sub_and_test - subtract value from variable and test result
1277
- * @i: integer value to subtract
1278
- * @v: pointer of type atomic_t
1279
- *
1280
- * Atomically subtracts @i from @v and returns
1281
- * true if the result is zero, or false for all
1282
- * other cases.
1283
- */
1284
1275
static __always_inline bool
1285
1276
arch_atomic_sub_and_test (int i , atomic_t * v )
1286
1277
{
@@ -1290,14 +1281,6 @@ arch_atomic_sub_and_test(int i, atomic_t *v)
1290
1281
#endif
1291
1282
1292
1283
#ifndef arch_atomic_dec_and_test
1293
- /**
1294
- * arch_atomic_dec_and_test - decrement and test
1295
- * @v: pointer of type atomic_t
1296
- *
1297
- * Atomically decrements @v by 1 and
1298
- * returns true if the result is 0, or false for all other
1299
- * cases.
1300
- */
1301
1284
static __always_inline bool
1302
1285
arch_atomic_dec_and_test (atomic_t * v )
1303
1286
{
@@ -1307,14 +1290,6 @@ arch_atomic_dec_and_test(atomic_t *v)
1307
1290
#endif
1308
1291
1309
1292
#ifndef arch_atomic_inc_and_test
1310
- /**
1311
- * arch_atomic_inc_and_test - increment and test
1312
- * @v: pointer of type atomic_t
1313
- *
1314
- * Atomically increments @v by 1
1315
- * and returns true if the result is zero, or false for all
1316
- * other cases.
1317
- */
1318
1293
static __always_inline bool
1319
1294
arch_atomic_inc_and_test (atomic_t * v )
1320
1295
{
@@ -1331,14 +1306,6 @@ arch_atomic_inc_and_test(atomic_t *v)
1331
1306
#endif /* arch_atomic_add_negative */
1332
1307
1333
1308
#ifndef arch_atomic_add_negative
1334
- /**
1335
- * arch_atomic_add_negative - Add and test if negative
1336
- * @i: integer value to add
1337
- * @v: pointer of type atomic_t
1338
- *
1339
- * Atomically adds @i to @v and returns true if the result is negative,
1340
- * or false when the result is greater than or equal to zero.
1341
- */
1342
1309
static __always_inline bool
1343
1310
arch_atomic_add_negative (int i , atomic_t * v )
1344
1311
{
@@ -1348,14 +1315,6 @@ arch_atomic_add_negative(int i, atomic_t *v)
1348
1315
#endif
1349
1316
1350
1317
#ifndef arch_atomic_add_negative_acquire
1351
- /**
1352
- * arch_atomic_add_negative_acquire - Add and test if negative
1353
- * @i: integer value to add
1354
- * @v: pointer of type atomic_t
1355
- *
1356
- * Atomically adds @i to @v and returns true if the result is negative,
1357
- * or false when the result is greater than or equal to zero.
1358
- */
1359
1318
static __always_inline bool
1360
1319
arch_atomic_add_negative_acquire (int i , atomic_t * v )
1361
1320
{
@@ -1365,14 +1324,6 @@ arch_atomic_add_negative_acquire(int i, atomic_t *v)
1365
1324
#endif
1366
1325
1367
1326
#ifndef arch_atomic_add_negative_release
1368
- /**
1369
- * arch_atomic_add_negative_release - Add and test if negative
1370
- * @i: integer value to add
1371
- * @v: pointer of type atomic_t
1372
- *
1373
- * Atomically adds @i to @v and returns true if the result is negative,
1374
- * or false when the result is greater than or equal to zero.
1375
- */
1376
1327
static __always_inline bool
1377
1328
arch_atomic_add_negative_release (int i , atomic_t * v )
1378
1329
{
@@ -1382,14 +1333,6 @@ arch_atomic_add_negative_release(int i, atomic_t *v)
1382
1333
#endif
1383
1334
1384
1335
#ifndef arch_atomic_add_negative_relaxed
1385
- /**
1386
- * arch_atomic_add_negative_relaxed - Add and test if negative
1387
- * @i: integer value to add
1388
- * @v: pointer of type atomic_t
1389
- *
1390
- * Atomically adds @i to @v and returns true if the result is negative,
1391
- * or false when the result is greater than or equal to zero.
1392
- */
1393
1336
static __always_inline bool
1394
1337
arch_atomic_add_negative_relaxed (int i , atomic_t * v )
1395
1338
{
@@ -1437,15 +1380,6 @@ arch_atomic_add_negative(int i, atomic_t *v)
1437
1380
#endif /* arch_atomic_add_negative_relaxed */
1438
1381
1439
1382
#ifndef arch_atomic_fetch_add_unless
1440
- /**
1441
- * arch_atomic_fetch_add_unless - add unless the number is already a given value
1442
- * @v: pointer of type atomic_t
1443
- * @a: the amount to add to v...
1444
- * @u: ...unless v is equal to u.
1445
- *
1446
- * Atomically adds @a to @v, so long as @v was not already @u.
1447
- * Returns original value of @v
1448
- */
1449
1383
static __always_inline int
1450
1384
arch_atomic_fetch_add_unless (atomic_t * v , int a , int u )
1451
1385
{
@@ -1462,15 +1396,6 @@ arch_atomic_fetch_add_unless(atomic_t *v, int a, int u)
1462
1396
#endif
1463
1397
1464
1398
#ifndef arch_atomic_add_unless
1465
- /**
1466
- * arch_atomic_add_unless - add unless the number is already a given value
1467
- * @v: pointer of type atomic_t
1468
- * @a: the amount to add to v...
1469
- * @u: ...unless v is equal to u.
1470
- *
1471
- * Atomically adds @a to @v, if @v was not already @u.
1472
- * Returns true if the addition was done.
1473
- */
1474
1399
static __always_inline bool
1475
1400
arch_atomic_add_unless (atomic_t * v , int a , int u )
1476
1401
{
@@ -1480,13 +1405,6 @@ arch_atomic_add_unless(atomic_t *v, int a, int u)
1480
1405
#endif
1481
1406
1482
1407
#ifndef arch_atomic_inc_not_zero
1483
- /**
1484
- * arch_atomic_inc_not_zero - increment unless the number is zero
1485
- * @v: pointer of type atomic_t
1486
- *
1487
- * Atomically increments @v by 1, if @v is non-zero.
1488
- * Returns true if the increment was done.
1489
- */
1490
1408
static __always_inline bool
1491
1409
arch_atomic_inc_not_zero (atomic_t * v )
1492
1410
{
@@ -2488,15 +2406,6 @@ arch_atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new)
2488
2406
#endif /* arch_atomic64_try_cmpxchg_relaxed */
2489
2407
2490
2408
#ifndef arch_atomic64_sub_and_test
2491
- /**
2492
- * arch_atomic64_sub_and_test - subtract value from variable and test result
2493
- * @i: integer value to subtract
2494
- * @v: pointer of type atomic64_t
2495
- *
2496
- * Atomically subtracts @i from @v and returns
2497
- * true if the result is zero, or false for all
2498
- * other cases.
2499
- */
2500
2409
static __always_inline bool
2501
2410
arch_atomic64_sub_and_test (s64 i , atomic64_t * v )
2502
2411
{
@@ -2506,14 +2415,6 @@ arch_atomic64_sub_and_test(s64 i, atomic64_t *v)
2506
2415
#endif
2507
2416
2508
2417
#ifndef arch_atomic64_dec_and_test
2509
- /**
2510
- * arch_atomic64_dec_and_test - decrement and test
2511
- * @v: pointer of type atomic64_t
2512
- *
2513
- * Atomically decrements @v by 1 and
2514
- * returns true if the result is 0, or false for all other
2515
- * cases.
2516
- */
2517
2418
static __always_inline bool
2518
2419
arch_atomic64_dec_and_test (atomic64_t * v )
2519
2420
{
@@ -2523,14 +2424,6 @@ arch_atomic64_dec_and_test(atomic64_t *v)
2523
2424
#endif
2524
2425
2525
2426
#ifndef arch_atomic64_inc_and_test
2526
- /**
2527
- * arch_atomic64_inc_and_test - increment and test
2528
- * @v: pointer of type atomic64_t
2529
- *
2530
- * Atomically increments @v by 1
2531
- * and returns true if the result is zero, or false for all
2532
- * other cases.
2533
- */
2534
2427
static __always_inline bool
2535
2428
arch_atomic64_inc_and_test (atomic64_t * v )
2536
2429
{
@@ -2547,14 +2440,6 @@ arch_atomic64_inc_and_test(atomic64_t *v)
2547
2440
#endif /* arch_atomic64_add_negative */
2548
2441
2549
2442
#ifndef arch_atomic64_add_negative
2550
- /**
2551
- * arch_atomic64_add_negative - Add and test if negative
2552
- * @i: integer value to add
2553
- * @v: pointer of type atomic64_t
2554
- *
2555
- * Atomically adds @i to @v and returns true if the result is negative,
2556
- * or false when the result is greater than or equal to zero.
2557
- */
2558
2443
static __always_inline bool
2559
2444
arch_atomic64_add_negative (s64 i , atomic64_t * v )
2560
2445
{
@@ -2564,14 +2449,6 @@ arch_atomic64_add_negative(s64 i, atomic64_t *v)
2564
2449
#endif
2565
2450
2566
2451
#ifndef arch_atomic64_add_negative_acquire
2567
- /**
2568
- * arch_atomic64_add_negative_acquire - Add and test if negative
2569
- * @i: integer value to add
2570
- * @v: pointer of type atomic64_t
2571
- *
2572
- * Atomically adds @i to @v and returns true if the result is negative,
2573
- * or false when the result is greater than or equal to zero.
2574
- */
2575
2452
static __always_inline bool
2576
2453
arch_atomic64_add_negative_acquire (s64 i , atomic64_t * v )
2577
2454
{
@@ -2581,14 +2458,6 @@ arch_atomic64_add_negative_acquire(s64 i, atomic64_t *v)
2581
2458
#endif
2582
2459
2583
2460
#ifndef arch_atomic64_add_negative_release
2584
- /**
2585
- * arch_atomic64_add_negative_release - Add and test if negative
2586
- * @i: integer value to add
2587
- * @v: pointer of type atomic64_t
2588
- *
2589
- * Atomically adds @i to @v and returns true if the result is negative,
2590
- * or false when the result is greater than or equal to zero.
2591
- */
2592
2461
static __always_inline bool
2593
2462
arch_atomic64_add_negative_release (s64 i , atomic64_t * v )
2594
2463
{
@@ -2598,14 +2467,6 @@ arch_atomic64_add_negative_release(s64 i, atomic64_t *v)
2598
2467
#endif
2599
2468
2600
2469
#ifndef arch_atomic64_add_negative_relaxed
2601
- /**
2602
- * arch_atomic64_add_negative_relaxed - Add and test if negative
2603
- * @i: integer value to add
2604
- * @v: pointer of type atomic64_t
2605
- *
2606
- * Atomically adds @i to @v and returns true if the result is negative,
2607
- * or false when the result is greater than or equal to zero.
2608
- */
2609
2470
static __always_inline bool
2610
2471
arch_atomic64_add_negative_relaxed (s64 i , atomic64_t * v )
2611
2472
{
@@ -2653,15 +2514,6 @@ arch_atomic64_add_negative(s64 i, atomic64_t *v)
2653
2514
#endif /* arch_atomic64_add_negative_relaxed */
2654
2515
2655
2516
#ifndef arch_atomic64_fetch_add_unless
2656
- /**
2657
- * arch_atomic64_fetch_add_unless - add unless the number is already a given value
2658
- * @v: pointer of type atomic64_t
2659
- * @a: the amount to add to v...
2660
- * @u: ...unless v is equal to u.
2661
- *
2662
- * Atomically adds @a to @v, so long as @v was not already @u.
2663
- * Returns original value of @v
2664
- */
2665
2517
static __always_inline s64
2666
2518
arch_atomic64_fetch_add_unless (atomic64_t * v , s64 a , s64 u )
2667
2519
{
@@ -2678,15 +2530,6 @@ arch_atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
2678
2530
#endif
2679
2531
2680
2532
#ifndef arch_atomic64_add_unless
2681
- /**
2682
- * arch_atomic64_add_unless - add unless the number is already a given value
2683
- * @v: pointer of type atomic64_t
2684
- * @a: the amount to add to v...
2685
- * @u: ...unless v is equal to u.
2686
- *
2687
- * Atomically adds @a to @v, if @v was not already @u.
2688
- * Returns true if the addition was done.
2689
- */
2690
2533
static __always_inline bool
2691
2534
arch_atomic64_add_unless (atomic64_t * v , s64 a , s64 u )
2692
2535
{
@@ -2696,13 +2539,6 @@ arch_atomic64_add_unless(atomic64_t *v, s64 a, s64 u)
2696
2539
#endif
2697
2540
2698
2541
#ifndef arch_atomic64_inc_not_zero
2699
- /**
2700
- * arch_atomic64_inc_not_zero - increment unless the number is zero
2701
- * @v: pointer of type atomic64_t
2702
- *
2703
- * Atomically increments @v by 1, if @v is non-zero.
2704
- * Returns true if the increment was done.
2705
- */
2706
2542
static __always_inline bool
2707
2543
arch_atomic64_inc_not_zero (atomic64_t * v )
2708
2544
{
@@ -2761,4 +2597,4 @@ arch_atomic64_dec_if_positive(atomic64_t *v)
2761
2597
#endif
2762
2598
2763
2599
#endif /* _LINUX_ATOMIC_FALLBACK_H */
2764
- // 52dfc6fe4a2e7234bbd2aa3e16a377c1db793a53
2600
+ // 9f0fd6ed53267c6ec64e36cd18e6fd8df57ea277
0 commit comments