|
@@ -82,7 +82,7 @@ static int aes_set_key_common(struct crypto_tfm *tfm, void *raw_ctx,
|
|
|
return -EINVAL;
|
|
|
}
|
|
|
|
|
|
- if (irq_fpu_usable())
|
|
|
+ if (!irq_fpu_usable())
|
|
|
err = crypto_aes_expand_key(ctx, in_key, key_len);
|
|
|
else {
|
|
|
kernel_fpu_begin();
|
|
@@ -103,7 +103,7 @@ static void aes_encrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
|
|
{
|
|
|
struct crypto_aes_ctx *ctx = aes_ctx(crypto_tfm_ctx(tfm));
|
|
|
|
|
|
- if (irq_fpu_usable())
|
|
|
+ if (!irq_fpu_usable())
|
|
|
crypto_aes_encrypt_x86(ctx, dst, src);
|
|
|
else {
|
|
|
kernel_fpu_begin();
|
|
@@ -116,7 +116,7 @@ static void aes_decrypt(struct crypto_tfm *tfm, u8 *dst, const u8 *src)
|
|
|
{
|
|
|
struct crypto_aes_ctx *ctx = aes_ctx(crypto_tfm_ctx(tfm));
|
|
|
|
|
|
- if (irq_fpu_usable())
|
|
|
+ if (!irq_fpu_usable())
|
|
|
crypto_aes_decrypt_x86(ctx, dst, src);
|
|
|
else {
|
|
|
kernel_fpu_begin();
|
|
@@ -342,7 +342,7 @@ static int ablk_encrypt(struct ablkcipher_request *req)
|
|
|
struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
|
|
|
struct async_aes_ctx *ctx = crypto_ablkcipher_ctx(tfm);
|
|
|
|
|
|
- if (irq_fpu_usable()) {
|
|
|
+ if (!irq_fpu_usable()) {
|
|
|
struct ablkcipher_request *cryptd_req =
|
|
|
ablkcipher_request_ctx(req);
|
|
|
memcpy(cryptd_req, req, sizeof(*req));
|
|
@@ -363,7 +363,7 @@ static int ablk_decrypt(struct ablkcipher_request *req)
|
|
|
struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
|
|
|
struct async_aes_ctx *ctx = crypto_ablkcipher_ctx(tfm);
|
|
|
|
|
|
- if (irq_fpu_usable()) {
|
|
|
+ if (!irq_fpu_usable()) {
|
|
|
struct ablkcipher_request *cryptd_req =
|
|
|
ablkcipher_request_ctx(req);
|
|
|
memcpy(cryptd_req, req, sizeof(*req));
|