summaryrefslogtreecommitdiffstats
path: root/include/crypto
diff options
context:
space:
mode:
authorCorentin Labbe <clabbe@baylibre.com>2018-11-29 15:42:18 +0100
committerHerbert Xu <herbert@gondor.apana.org.au>2018-12-07 07:15:00 +0100
commit6e8e72cd206e2ba68801e4f2490f639d41808c8d (patch)
tree6c89421ccb76ab0abe12e0d132c4c9a6a9de914b /include/crypto
parentcrypto: user - CRYPTO_STATS should depend on CRYPTO_USER (diff)
downloadlinux-6e8e72cd206e2ba68801e4f2490f639d41808c8d.tar.xz
linux-6e8e72cd206e2ba68801e4f2490f639d41808c8d.zip
crypto: user - convert all stats from u32 to u64
All the 32-bit fields need to be 64-bit. In some cases, UINT32_MAX crypto operations can be done in seconds. Reported-by: Eric Biggers <ebiggers@kernel.org> Signed-off-by: Corentin Labbe <clabbe@baylibre.com> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
Diffstat (limited to 'include/crypto')
-rw-r--r--include/crypto/acompress.h8
-rw-r--r--include/crypto/aead.h8
-rw-r--r--include/crypto/akcipher.h16
-rw-r--r--include/crypto/hash.h6
-rw-r--r--include/crypto/kpp.h12
-rw-r--r--include/crypto/rng.h8
-rw-r--r--include/crypto/skcipher.h8
7 files changed, 33 insertions, 33 deletions
diff --git a/include/crypto/acompress.h b/include/crypto/acompress.h
index 22e6f412c595..f79918196811 100644
--- a/include/crypto/acompress.h
+++ b/include/crypto/acompress.h
@@ -240,9 +240,9 @@ static inline void crypto_stat_compress(struct acomp_req *req, int ret)
struct crypto_acomp *tfm = crypto_acomp_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->compress_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->compress_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->compress_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->compress_cnt);
atomic64_add(req->slen, &tfm->base.__crt_alg->compress_tlen);
}
#endif
@@ -254,9 +254,9 @@ static inline void crypto_stat_decompress(struct acomp_req *req, int ret)
struct crypto_acomp *tfm = crypto_acomp_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->compress_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->compress_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->decompress_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->decompress_cnt);
atomic64_add(req->slen, &tfm->base.__crt_alg->decompress_tlen);
}
#endif
diff --git a/include/crypto/aead.h b/include/crypto/aead.h
index 0d765d7bfb82..99afd78c665d 100644
--- a/include/crypto/aead.h
+++ b/include/crypto/aead.h
@@ -312,9 +312,9 @@ static inline void crypto_stat_aead_encrypt(struct aead_request *req, int ret)
struct crypto_aead *tfm = crypto_aead_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->aead_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->aead_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->encrypt_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->encrypt_cnt);
atomic64_add(req->cryptlen, &tfm->base.__crt_alg->encrypt_tlen);
}
#endif
@@ -326,9 +326,9 @@ static inline void crypto_stat_aead_decrypt(struct aead_request *req, int ret)
struct crypto_aead *tfm = crypto_aead_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->aead_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->aead_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->decrypt_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->decrypt_cnt);
atomic64_add(req->cryptlen, &tfm->base.__crt_alg->decrypt_tlen);
}
#endif
diff --git a/include/crypto/akcipher.h b/include/crypto/akcipher.h
index afac71119396..3dc05cf7e0a9 100644
--- a/include/crypto/akcipher.h
+++ b/include/crypto/akcipher.h
@@ -278,9 +278,9 @@ static inline void crypto_stat_akcipher_encrypt(struct akcipher_request *req,
struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->encrypt_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->encrypt_cnt);
atomic64_add(req->src_len, &tfm->base.__crt_alg->encrypt_tlen);
}
#endif
@@ -293,9 +293,9 @@ static inline void crypto_stat_akcipher_decrypt(struct akcipher_request *req,
struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->decrypt_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->decrypt_cnt);
atomic64_add(req->src_len, &tfm->base.__crt_alg->decrypt_tlen);
}
#endif
@@ -308,9 +308,9 @@ static inline void crypto_stat_akcipher_sign(struct akcipher_request *req,
struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY)
- atomic_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
else
- atomic_inc(&tfm->base.__crt_alg->sign_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->sign_cnt);
#endif
}
@@ -321,9 +321,9 @@ static inline void crypto_stat_akcipher_verify(struct akcipher_request *req,
struct crypto_akcipher *tfm = crypto_akcipher_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY)
- atomic_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->akcipher_err_cnt);
else
- atomic_inc(&tfm->base.__crt_alg->verify_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->verify_cnt);
#endif
}
diff --git a/include/crypto/hash.h b/include/crypto/hash.h
index bc7796600338..52920bed05ba 100644
--- a/include/crypto/hash.h
+++ b/include/crypto/hash.h
@@ -418,7 +418,7 @@ static inline void crypto_stat_ahash_update(struct ahash_request *req, int ret)
struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY)
- atomic_inc(&tfm->base.__crt_alg->hash_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->hash_err_cnt);
else
atomic64_add(req->nbytes, &tfm->base.__crt_alg->hash_tlen);
#endif
@@ -430,9 +430,9 @@ static inline void crypto_stat_ahash_final(struct ahash_request *req, int ret)
struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->hash_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->hash_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->hash_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->hash_cnt);
atomic64_add(req->nbytes, &tfm->base.__crt_alg->hash_tlen);
}
#endif
diff --git a/include/crypto/kpp.h b/include/crypto/kpp.h
index f517ba6d3a27..bd5103a80919 100644
--- a/include/crypto/kpp.h
+++ b/include/crypto/kpp.h
@@ -272,9 +272,9 @@ static inline void crypto_stat_kpp_set_secret(struct crypto_kpp *tfm, int ret)
{
#ifdef CONFIG_CRYPTO_STATS
if (ret)
- atomic_inc(&tfm->base.__crt_alg->kpp_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->kpp_err_cnt);
else
- atomic_inc(&tfm->base.__crt_alg->setsecret_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->setsecret_cnt);
#endif
}
@@ -285,9 +285,9 @@ static inline void crypto_stat_kpp_generate_public_key(struct kpp_request *req,
struct crypto_kpp *tfm = crypto_kpp_reqtfm(req);
if (ret)
- atomic_inc(&tfm->base.__crt_alg->kpp_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->kpp_err_cnt);
else
- atomic_inc(&tfm->base.__crt_alg->generate_public_key_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->generate_public_key_cnt);
#endif
}
@@ -298,9 +298,9 @@ static inline void crypto_stat_kpp_compute_shared_secret(struct kpp_request *req
struct crypto_kpp *tfm = crypto_kpp_reqtfm(req);
if (ret)
- atomic_inc(&tfm->base.__crt_alg->kpp_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->kpp_err_cnt);
else
- atomic_inc(&tfm->base.__crt_alg->compute_shared_secret_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->compute_shared_secret_cnt);
#endif
}
diff --git a/include/crypto/rng.h b/include/crypto/rng.h
index 6d258f5b68f1..966615bba45e 100644
--- a/include/crypto/rng.h
+++ b/include/crypto/rng.h
@@ -126,9 +126,9 @@ static inline void crypto_stat_rng_seed(struct crypto_rng *tfm, int ret)
{
#ifdef CONFIG_CRYPTO_STATS
if (ret && ret != -EINPROGRESS && ret != -EBUSY)
- atomic_inc(&tfm->base.__crt_alg->rng_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->rng_err_cnt);
else
- atomic_inc(&tfm->base.__crt_alg->seed_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->seed_cnt);
#endif
}
@@ -137,9 +137,9 @@ static inline void crypto_stat_rng_generate(struct crypto_rng *tfm,
{
#ifdef CONFIG_CRYPTO_STATS
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&tfm->base.__crt_alg->rng_err_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->rng_err_cnt);
} else {
- atomic_inc(&tfm->base.__crt_alg->generate_cnt);
+ atomic64_inc(&tfm->base.__crt_alg->generate_cnt);
atomic64_add(dlen, &tfm->base.__crt_alg->generate_tlen);
}
#endif
diff --git a/include/crypto/skcipher.h b/include/crypto/skcipher.h
index 925f547cdcfa..dff54731ddf4 100644
--- a/include/crypto/skcipher.h
+++ b/include/crypto/skcipher.h
@@ -491,9 +491,9 @@ static inline void crypto_stat_skcipher_encrypt(struct skcipher_request *req,
{
#ifdef CONFIG_CRYPTO_STATS
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&alg->cipher_err_cnt);
+ atomic64_inc(&alg->cipher_err_cnt);
} else {
- atomic_inc(&alg->encrypt_cnt);
+ atomic64_inc(&alg->encrypt_cnt);
atomic64_add(req->cryptlen, &alg->encrypt_tlen);
}
#endif
@@ -504,9 +504,9 @@ static inline void crypto_stat_skcipher_decrypt(struct skcipher_request *req,
{
#ifdef CONFIG_CRYPTO_STATS
if (ret && ret != -EINPROGRESS && ret != -EBUSY) {
- atomic_inc(&alg->cipher_err_cnt);
+ atomic64_inc(&alg->cipher_err_cnt);
} else {
- atomic_inc(&alg->decrypt_cnt);
+ atomic64_inc(&alg->decrypt_cnt);
atomic64_add(req->cryptlen, &alg->decrypt_tlen);
}
#endif