@@ -523,25 +523,25 @@ zfs_zevent_post(nvlist_t *nvl, nvlist_t *detector, zevent_cb_t *cb)
523
523
524
524
error = nvlist_add_int64_array (nvl , FM_EREPORT_TIME , tv_array , 2 );
525
525
if (error ) {
526
- atomic_add_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 , 1 );
526
+ atomic_inc_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 );
527
527
goto out ;
528
528
}
529
529
530
530
eid = atomic_inc_64_nv (& zevent_eid );
531
531
error = nvlist_add_uint64 (nvl , FM_EREPORT_EID , eid );
532
532
if (error ) {
533
- atomic_add_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 , 1 );
533
+ atomic_inc_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 );
534
534
goto out ;
535
535
}
536
536
537
537
error = nvlist_size (nvl , & nvl_size , NV_ENCODE_NATIVE );
538
538
if (error ) {
539
- atomic_add_64 (& erpt_kstat_data .erpt_dropped .value .ui64 , 1 );
539
+ atomic_inc_64 (& erpt_kstat_data .erpt_dropped .value .ui64 );
540
540
goto out ;
541
541
}
542
542
543
543
if (nvl_size > ERPT_DATA_SZ || nvl_size == 0 ) {
544
- atomic_add_64 (& erpt_kstat_data .erpt_dropped .value .ui64 , 1 );
544
+ atomic_inc_64 (& erpt_kstat_data .erpt_dropped .value .ui64 );
545
545
error = EOVERFLOW ;
546
546
goto out ;
547
547
}
@@ -551,7 +551,7 @@ zfs_zevent_post(nvlist_t *nvl, nvlist_t *detector, zevent_cb_t *cb)
551
551
552
552
ev = zfs_zevent_alloc ();
553
553
if (ev == NULL ) {
554
- atomic_add_64 (& erpt_kstat_data .erpt_dropped .value .ui64 , 1 );
554
+ atomic_inc_64 (& erpt_kstat_data .erpt_dropped .value .ui64 );
555
555
error = ENOMEM ;
556
556
goto out ;
557
557
}
@@ -1025,8 +1025,7 @@ fm_payload_set(nvlist_t *payload, ...)
1025
1025
va_end (ap );
1026
1026
1027
1027
if (ret )
1028
- atomic_add_64 (
1029
- & erpt_kstat_data .payload_set_failed .value .ui64 , 1 );
1028
+ atomic_inc_64 (& erpt_kstat_data .payload_set_failed .value .ui64 );
1030
1029
}
1031
1030
1032
1031
/*
@@ -1059,24 +1058,24 @@ fm_ereport_set(nvlist_t *ereport, int version, const char *erpt_class,
1059
1058
int ret ;
1060
1059
1061
1060
if (version != FM_EREPORT_VERS0 ) {
1062
- atomic_add_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 , 1 );
1061
+ atomic_inc_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 );
1063
1062
return ;
1064
1063
}
1065
1064
1066
1065
(void ) snprintf (ereport_class , FM_MAX_CLASS , "%s.%s" ,
1067
1066
FM_EREPORT_CLASS , erpt_class );
1068
1067
if (nvlist_add_string (ereport , FM_CLASS , ereport_class ) != 0 ) {
1069
- atomic_add_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 , 1 );
1068
+ atomic_inc_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 );
1070
1069
return ;
1071
1070
}
1072
1071
1073
1072
if (nvlist_add_uint64 (ereport , FM_EREPORT_ENA , ena )) {
1074
- atomic_add_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 , 1 );
1073
+ atomic_inc_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 );
1075
1074
}
1076
1075
1077
1076
if (nvlist_add_nvlist (ereport , FM_EREPORT_DETECTOR ,
1078
1077
(nvlist_t * )detector ) != 0 ) {
1079
- atomic_add_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 , 1 );
1078
+ atomic_inc_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 );
1080
1079
}
1081
1080
1082
1081
va_start (ap , detector );
@@ -1085,7 +1084,7 @@ fm_ereport_set(nvlist_t *ereport, int version, const char *erpt_class,
1085
1084
va_end (ap );
1086
1085
1087
1086
if (ret )
1088
- atomic_add_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 , 1 );
1087
+ atomic_inc_64 (& erpt_kstat_data .erpt_set_failed .value .ui64 );
1089
1088
}
1090
1089
1091
1090
/*
@@ -1108,19 +1107,19 @@ static int
1108
1107
fm_fmri_hc_set_common (nvlist_t * fmri , int version , const nvlist_t * auth )
1109
1108
{
1110
1109
if (version != FM_HC_SCHEME_VERSION ) {
1111
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1110
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1112
1111
return (0 );
1113
1112
}
1114
1113
1115
1114
if (nvlist_add_uint8 (fmri , FM_VERSION , version ) != 0 ||
1116
1115
nvlist_add_string (fmri , FM_FMRI_SCHEME , FM_FMRI_SCHEME_HC ) != 0 ) {
1117
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1116
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1118
1117
return (0 );
1119
1118
}
1120
1119
1121
1120
if (auth != NULL && nvlist_add_nvlist (fmri , FM_FMRI_AUTHORITY ,
1122
1121
(nvlist_t * )auth ) != 0 ) {
1123
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1122
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1124
1123
return (0 );
1125
1124
}
1126
1125
@@ -1152,22 +1151,22 @@ fm_fmri_hc_set(nvlist_t *fmri, int version, const nvlist_t *auth,
1152
1151
pairs [i ] = fm_nvlist_create (nva );
1153
1152
if (nvlist_add_string (pairs [i ], FM_FMRI_HC_NAME , name ) != 0 ||
1154
1153
nvlist_add_string (pairs [i ], FM_FMRI_HC_ID , idstr ) != 0 ) {
1155
- atomic_add_64 (
1156
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1154
+ atomic_inc_64 (
1155
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1157
1156
}
1158
1157
}
1159
1158
va_end (ap );
1160
1159
1161
1160
if (nvlist_add_nvlist_array (fmri , FM_FMRI_HC_LIST , pairs , npairs ) != 0 )
1162
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1161
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1163
1162
1164
1163
for (i = 0 ; i < npairs ; i ++ )
1165
1164
fm_nvlist_destroy (pairs [i ], FM_NVA_RETAIN );
1166
1165
1167
1166
if (snvl != NULL ) {
1168
1167
if (nvlist_add_nvlist (fmri , FM_FMRI_HC_SPECIFIC , snvl ) != 0 ) {
1169
- atomic_add_64 (
1170
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1168
+ atomic_inc_64 (
1169
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1171
1170
}
1172
1171
}
1173
1172
}
@@ -1192,20 +1191,20 @@ fm_fmri_hc_create(nvlist_t *fmri, int version, const nvlist_t *auth,
1192
1191
*/
1193
1192
if (nvlist_lookup_nvlist_array (bboard , FM_FMRI_HC_LIST , & hcl , & n )
1194
1193
!= 0 ) {
1195
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1194
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1196
1195
return ;
1197
1196
}
1198
1197
1199
1198
for (i = 0 ; i < n ; i ++ ) {
1200
1199
if (nvlist_lookup_string (hcl [i ], FM_FMRI_HC_NAME ,
1201
1200
& hcname ) != 0 ) {
1202
- atomic_add_64 (
1203
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1201
+ atomic_inc_64 (
1202
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1204
1203
return ;
1205
1204
}
1206
1205
if (nvlist_lookup_string (hcl [i ], FM_FMRI_HC_ID , & hcid ) != 0 ) {
1207
- atomic_add_64 (
1208
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1206
+ atomic_inc_64 (
1207
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1209
1208
return ;
1210
1209
}
1211
1210
@@ -1217,8 +1216,8 @@ fm_fmri_hc_create(nvlist_t *fmri, int version, const nvlist_t *auth,
1217
1216
fm_nvlist_destroy (pairs [j ],
1218
1217
FM_NVA_RETAIN );
1219
1218
}
1220
- atomic_add_64 (
1221
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1219
+ atomic_inc_64 (
1220
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1222
1221
return ;
1223
1222
}
1224
1223
}
@@ -1242,8 +1241,8 @@ fm_fmri_hc_create(nvlist_t *fmri, int version, const nvlist_t *auth,
1242
1241
fm_nvlist_destroy (pairs [j ],
1243
1242
FM_NVA_RETAIN );
1244
1243
}
1245
- atomic_add_64 (
1246
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1244
+ atomic_inc_64 (
1245
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1247
1246
return ;
1248
1247
}
1249
1248
}
@@ -1254,7 +1253,7 @@ fm_fmri_hc_create(nvlist_t *fmri, int version, const nvlist_t *auth,
1254
1253
*/
1255
1254
if (nvlist_add_nvlist_array (fmri , FM_FMRI_HC_LIST , pairs ,
1256
1255
npairs + n ) != 0 ) {
1257
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1256
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1258
1257
return ;
1259
1258
}
1260
1259
@@ -1264,8 +1263,8 @@ fm_fmri_hc_create(nvlist_t *fmri, int version, const nvlist_t *auth,
1264
1263
1265
1264
if (snvl != NULL ) {
1266
1265
if (nvlist_add_nvlist (fmri , FM_FMRI_HC_SPECIFIC , snvl ) != 0 ) {
1267
- atomic_add_64 (
1268
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1266
+ atomic_inc_64 (
1267
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1269
1268
return ;
1270
1269
}
1271
1270
}
@@ -1291,7 +1290,7 @@ fm_fmri_dev_set(nvlist_t *fmri_dev, int version, const nvlist_t *auth,
1291
1290
int err = 0 ;
1292
1291
1293
1292
if (version != DEV_SCHEME_VERSION0 ) {
1294
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1293
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1295
1294
return ;
1296
1295
}
1297
1296
@@ -1312,7 +1311,7 @@ fm_fmri_dev_set(nvlist_t *fmri_dev, int version, const nvlist_t *auth,
1312
1311
err |= nvlist_add_string (fmri_dev , FM_FMRI_DEV_TGTPTLUN0 , tpl0 );
1313
1312
1314
1313
if (err )
1315
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1314
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1316
1315
1317
1316
}
1318
1317
@@ -1337,35 +1336,35 @@ fm_fmri_cpu_set(nvlist_t *fmri_cpu, int version, const nvlist_t *auth,
1337
1336
uint64_t * failedp = & erpt_kstat_data .fmri_set_failed .value .ui64 ;
1338
1337
1339
1338
if (version < CPU_SCHEME_VERSION1 ) {
1340
- atomic_add_64 (failedp , 1 );
1339
+ atomic_inc_64 (failedp );
1341
1340
return ;
1342
1341
}
1343
1342
1344
1343
if (nvlist_add_uint8 (fmri_cpu , FM_VERSION , version ) != 0 ) {
1345
- atomic_add_64 (failedp , 1 );
1344
+ atomic_inc_64 (failedp );
1346
1345
return ;
1347
1346
}
1348
1347
1349
1348
if (nvlist_add_string (fmri_cpu , FM_FMRI_SCHEME ,
1350
1349
FM_FMRI_SCHEME_CPU ) != 0 ) {
1351
- atomic_add_64 (failedp , 1 );
1350
+ atomic_inc_64 (failedp );
1352
1351
return ;
1353
1352
}
1354
1353
1355
1354
if (auth != NULL && nvlist_add_nvlist (fmri_cpu , FM_FMRI_AUTHORITY ,
1356
1355
(nvlist_t * )auth ) != 0 )
1357
- atomic_add_64 (failedp , 1 );
1356
+ atomic_inc_64 (failedp );
1358
1357
1359
1358
if (nvlist_add_uint32 (fmri_cpu , FM_FMRI_CPU_ID , cpu_id ) != 0 )
1360
- atomic_add_64 (failedp , 1 );
1359
+ atomic_inc_64 (failedp );
1361
1360
1362
1361
if (cpu_maskp != NULL && nvlist_add_uint8 (fmri_cpu , FM_FMRI_CPU_MASK ,
1363
1362
* cpu_maskp ) != 0 )
1364
- atomic_add_64 (failedp , 1 );
1363
+ atomic_inc_64 (failedp );
1365
1364
1366
1365
if (serial_idp == NULL || nvlist_add_string (fmri_cpu ,
1367
1366
FM_FMRI_CPU_SERIAL_ID , (char * )serial_idp ) != 0 )
1368
- atomic_add_64 (failedp , 1 );
1367
+ atomic_inc_64 (failedp );
1369
1368
}
1370
1369
1371
1370
/*
@@ -1386,49 +1385,47 @@ fm_fmri_mem_set(nvlist_t *fmri, int version, const nvlist_t *auth,
1386
1385
const char * unum , const char * serial , uint64_t offset )
1387
1386
{
1388
1387
if (version != MEM_SCHEME_VERSION0 ) {
1389
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1388
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1390
1389
return ;
1391
1390
}
1392
1391
1393
1392
if (!serial && (offset != (uint64_t )-1 )) {
1394
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1393
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1395
1394
return ;
1396
1395
}
1397
1396
1398
1397
if (nvlist_add_uint8 (fmri , FM_VERSION , version ) != 0 ) {
1399
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1398
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1400
1399
return ;
1401
1400
}
1402
1401
1403
1402
if (nvlist_add_string (fmri , FM_FMRI_SCHEME , FM_FMRI_SCHEME_MEM ) != 0 ) {
1404
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1403
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1405
1404
return ;
1406
1405
}
1407
1406
1408
1407
if (auth != NULL ) {
1409
1408
if (nvlist_add_nvlist (fmri , FM_FMRI_AUTHORITY ,
1410
1409
(nvlist_t * )auth ) != 0 ) {
1411
- atomic_add_64 (
1412
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1410
+ atomic_inc_64 (
1411
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1413
1412
}
1414
1413
}
1415
1414
1416
1415
if (nvlist_add_string (fmri , FM_FMRI_MEM_UNUM , unum ) != 0 ) {
1417
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1416
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1418
1417
}
1419
1418
1420
1419
if (serial != NULL ) {
1421
1420
if (nvlist_add_string_array (fmri , FM_FMRI_MEM_SERIAL_ID ,
1422
1421
(char * * )& serial , 1 ) != 0 ) {
1423
- atomic_add_64 (
1424
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1422
+ atomic_inc_64 (
1423
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1425
1424
}
1426
- if (offset != (uint64_t )-1 ) {
1427
- if (nvlist_add_uint64 (fmri , FM_FMRI_MEM_OFFSET ,
1428
- offset ) != 0 ) {
1429
- atomic_add_64 (& erpt_kstat_data .
1430
- fmri_set_failed .value .ui64 , 1 );
1431
- }
1425
+ if (offset != (uint64_t )-1 && nvlist_add_uint64 (fmri ,
1426
+ FM_FMRI_MEM_OFFSET , offset ) != 0 ) {
1427
+ atomic_inc_64 (
1428
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1432
1429
}
1433
1430
}
1434
1431
}
@@ -1438,28 +1435,28 @@ fm_fmri_zfs_set(nvlist_t *fmri, int version, uint64_t pool_guid,
1438
1435
uint64_t vdev_guid )
1439
1436
{
1440
1437
if (version != ZFS_SCHEME_VERSION0 ) {
1441
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1438
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1442
1439
return ;
1443
1440
}
1444
1441
1445
1442
if (nvlist_add_uint8 (fmri , FM_VERSION , version ) != 0 ) {
1446
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1443
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1447
1444
return ;
1448
1445
}
1449
1446
1450
1447
if (nvlist_add_string (fmri , FM_FMRI_SCHEME , FM_FMRI_SCHEME_ZFS ) != 0 ) {
1451
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1448
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1452
1449
return ;
1453
1450
}
1454
1451
1455
1452
if (nvlist_add_uint64 (fmri , FM_FMRI_ZFS_POOL , pool_guid ) != 0 ) {
1456
- atomic_add_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1453
+ atomic_inc_64 (& erpt_kstat_data .fmri_set_failed .value .ui64 );
1457
1454
}
1458
1455
1459
1456
if (vdev_guid != 0 ) {
1460
1457
if (nvlist_add_uint64 (fmri , FM_FMRI_ZFS_VDEV , vdev_guid ) != 0 ) {
1461
- atomic_add_64 (
1462
- & erpt_kstat_data .fmri_set_failed .value .ui64 , 1 );
1458
+ atomic_inc_64 (
1459
+ & erpt_kstat_data .fmri_set_failed .value .ui64 );
1463
1460
}
1464
1461
}
1465
1462
}
0 commit comments