serpent_sse2_glue.c 21 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841
  1. /*
  2. * Glue Code for SSE2 assembler versions of Serpent Cipher
  3. *
  4. * Copyright (c) 2011 Jussi Kivilinna <jussi.kivilinna@mbnet.fi>
  5. *
  6. * Glue code based on aesni-intel_glue.c by:
  7. * Copyright (C) 2008, Intel Corp.
  8. * Author: Huang Ying <ying.huang@intel.com>
  9. *
  10. * CBC & ECB parts based on code (crypto/cbc.c,ecb.c) by:
  11. * Copyright (c) 2006 Herbert Xu <herbert@gondor.apana.org.au>
  12. * CTR part based on code (crypto/ctr.c) by:
  13. * (C) Copyright IBM Corp. 2007 - Joy Latten <latten@us.ibm.com>
  14. *
  15. * This program is free software; you can redistribute it and/or modify
  16. * it under the terms of the GNU General Public License as published by
  17. * the Free Software Foundation; either version 2 of the License, or
  18. * (at your option) any later version.
  19. *
  20. * This program is distributed in the hope that it will be useful,
  21. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  22. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  23. * GNU General Public License for more details.
  24. *
  25. * You should have received a copy of the GNU General Public License
  26. * along with this program; if not, write to the Free Software
  27. * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
  28. * USA
  29. *
  30. */
  31. #include <linux/module.h>
  32. #include <linux/hardirq.h>
  33. #include <linux/types.h>
  34. #include <linux/crypto.h>
  35. #include <linux/err.h>
  36. #include <crypto/algapi.h>
  37. #include <crypto/serpent.h>
  38. #include <crypto/cryptd.h>
  39. #include <crypto/b128ops.h>
  40. #include <crypto/ctr.h>
  41. #include <crypto/lrw.h>
  42. #include <crypto/xts.h>
  43. #include <asm/i387.h>
  44. #include <asm/serpent-sse2.h>
  45. #include <asm/crypto/ablk_helper.h>
  46. #include <crypto/scatterwalk.h>
  47. #include <linux/workqueue.h>
  48. #include <linux/spinlock.h>
  49. static inline bool serpent_fpu_begin(bool fpu_enabled, unsigned int nbytes)
  50. {
  51. if (fpu_enabled)
  52. return true;
  53. /* SSE2 is only used when chunk to be processed is large enough, so
  54. * do not enable FPU until it is necessary.
  55. */
  56. if (nbytes < SERPENT_BLOCK_SIZE * SERPENT_PARALLEL_BLOCKS)
  57. return false;
  58. kernel_fpu_begin();
  59. return true;
  60. }
  61. static inline void serpent_fpu_end(bool fpu_enabled)
  62. {
  63. if (fpu_enabled)
  64. kernel_fpu_end();
  65. }
  66. static int ecb_crypt(struct blkcipher_desc *desc, struct blkcipher_walk *walk,
  67. bool enc)
  68. {
  69. bool fpu_enabled = false;
  70. struct serpent_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  71. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  72. unsigned int nbytes;
  73. int err;
  74. err = blkcipher_walk_virt(desc, walk);
  75. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  76. while ((nbytes = walk->nbytes)) {
  77. u8 *wsrc = walk->src.virt.addr;
  78. u8 *wdst = walk->dst.virt.addr;
  79. fpu_enabled = serpent_fpu_begin(fpu_enabled, nbytes);
  80. /* Process multi-block batch */
  81. if (nbytes >= bsize * SERPENT_PARALLEL_BLOCKS) {
  82. do {
  83. if (enc)
  84. serpent_enc_blk_xway(ctx, wdst, wsrc);
  85. else
  86. serpent_dec_blk_xway(ctx, wdst, wsrc);
  87. wsrc += bsize * SERPENT_PARALLEL_BLOCKS;
  88. wdst += bsize * SERPENT_PARALLEL_BLOCKS;
  89. nbytes -= bsize * SERPENT_PARALLEL_BLOCKS;
  90. } while (nbytes >= bsize * SERPENT_PARALLEL_BLOCKS);
  91. if (nbytes < bsize)
  92. goto done;
  93. }
  94. /* Handle leftovers */
  95. do {
  96. if (enc)
  97. __serpent_encrypt(ctx, wdst, wsrc);
  98. else
  99. __serpent_decrypt(ctx, wdst, wsrc);
  100. wsrc += bsize;
  101. wdst += bsize;
  102. nbytes -= bsize;
  103. } while (nbytes >= bsize);
  104. done:
  105. err = blkcipher_walk_done(desc, walk, nbytes);
  106. }
  107. serpent_fpu_end(fpu_enabled);
  108. return err;
  109. }
  110. static int ecb_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  111. struct scatterlist *src, unsigned int nbytes)
  112. {
  113. struct blkcipher_walk walk;
  114. blkcipher_walk_init(&walk, dst, src, nbytes);
  115. return ecb_crypt(desc, &walk, true);
  116. }
  117. static int ecb_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  118. struct scatterlist *src, unsigned int nbytes)
  119. {
  120. struct blkcipher_walk walk;
  121. blkcipher_walk_init(&walk, dst, src, nbytes);
  122. return ecb_crypt(desc, &walk, false);
  123. }
  124. static unsigned int __cbc_encrypt(struct blkcipher_desc *desc,
  125. struct blkcipher_walk *walk)
  126. {
  127. struct serpent_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  128. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  129. unsigned int nbytes = walk->nbytes;
  130. u128 *src = (u128 *)walk->src.virt.addr;
  131. u128 *dst = (u128 *)walk->dst.virt.addr;
  132. u128 *iv = (u128 *)walk->iv;
  133. do {
  134. u128_xor(dst, src, iv);
  135. __serpent_encrypt(ctx, (u8 *)dst, (u8 *)dst);
  136. iv = dst;
  137. src += 1;
  138. dst += 1;
  139. nbytes -= bsize;
  140. } while (nbytes >= bsize);
  141. u128_xor((u128 *)walk->iv, (u128 *)walk->iv, iv);
  142. return nbytes;
  143. }
  144. static int cbc_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  145. struct scatterlist *src, unsigned int nbytes)
  146. {
  147. struct blkcipher_walk walk;
  148. int err;
  149. blkcipher_walk_init(&walk, dst, src, nbytes);
  150. err = blkcipher_walk_virt(desc, &walk);
  151. while ((nbytes = walk.nbytes)) {
  152. nbytes = __cbc_encrypt(desc, &walk);
  153. err = blkcipher_walk_done(desc, &walk, nbytes);
  154. }
  155. return err;
  156. }
  157. static unsigned int __cbc_decrypt(struct blkcipher_desc *desc,
  158. struct blkcipher_walk *walk)
  159. {
  160. struct serpent_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  161. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  162. unsigned int nbytes = walk->nbytes;
  163. u128 *src = (u128 *)walk->src.virt.addr;
  164. u128 *dst = (u128 *)walk->dst.virt.addr;
  165. u128 ivs[SERPENT_PARALLEL_BLOCKS - 1];
  166. u128 last_iv;
  167. int i;
  168. /* Start of the last block. */
  169. src += nbytes / bsize - 1;
  170. dst += nbytes / bsize - 1;
  171. last_iv = *src;
  172. /* Process multi-block batch */
  173. if (nbytes >= bsize * SERPENT_PARALLEL_BLOCKS) {
  174. do {
  175. nbytes -= bsize * (SERPENT_PARALLEL_BLOCKS - 1);
  176. src -= SERPENT_PARALLEL_BLOCKS - 1;
  177. dst -= SERPENT_PARALLEL_BLOCKS - 1;
  178. for (i = 0; i < SERPENT_PARALLEL_BLOCKS - 1; i++)
  179. ivs[i] = src[i];
  180. serpent_dec_blk_xway(ctx, (u8 *)dst, (u8 *)src);
  181. for (i = 0; i < SERPENT_PARALLEL_BLOCKS - 1; i++)
  182. u128_xor(dst + (i + 1), dst + (i + 1), ivs + i);
  183. nbytes -= bsize;
  184. if (nbytes < bsize)
  185. goto done;
  186. u128_xor(dst, dst, src - 1);
  187. src -= 1;
  188. dst -= 1;
  189. } while (nbytes >= bsize * SERPENT_PARALLEL_BLOCKS);
  190. if (nbytes < bsize)
  191. goto done;
  192. }
  193. /* Handle leftovers */
  194. for (;;) {
  195. __serpent_decrypt(ctx, (u8 *)dst, (u8 *)src);
  196. nbytes -= bsize;
  197. if (nbytes < bsize)
  198. break;
  199. u128_xor(dst, dst, src - 1);
  200. src -= 1;
  201. dst -= 1;
  202. }
  203. done:
  204. u128_xor(dst, dst, (u128 *)walk->iv);
  205. *(u128 *)walk->iv = last_iv;
  206. return nbytes;
  207. }
  208. static int cbc_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  209. struct scatterlist *src, unsigned int nbytes)
  210. {
  211. bool fpu_enabled = false;
  212. struct blkcipher_walk walk;
  213. int err;
  214. blkcipher_walk_init(&walk, dst, src, nbytes);
  215. err = blkcipher_walk_virt(desc, &walk);
  216. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  217. while ((nbytes = walk.nbytes)) {
  218. fpu_enabled = serpent_fpu_begin(fpu_enabled, nbytes);
  219. nbytes = __cbc_decrypt(desc, &walk);
  220. err = blkcipher_walk_done(desc, &walk, nbytes);
  221. }
  222. serpent_fpu_end(fpu_enabled);
  223. return err;
  224. }
  225. static inline void u128_to_be128(be128 *dst, const u128 *src)
  226. {
  227. dst->a = cpu_to_be64(src->a);
  228. dst->b = cpu_to_be64(src->b);
  229. }
  230. static inline void be128_to_u128(u128 *dst, const be128 *src)
  231. {
  232. dst->a = be64_to_cpu(src->a);
  233. dst->b = be64_to_cpu(src->b);
  234. }
  235. static inline void u128_inc(u128 *i)
  236. {
  237. i->b++;
  238. if (!i->b)
  239. i->a++;
  240. }
  241. static void ctr_crypt_final(struct blkcipher_desc *desc,
  242. struct blkcipher_walk *walk)
  243. {
  244. struct serpent_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  245. u8 *ctrblk = walk->iv;
  246. u8 keystream[SERPENT_BLOCK_SIZE];
  247. u8 *src = walk->src.virt.addr;
  248. u8 *dst = walk->dst.virt.addr;
  249. unsigned int nbytes = walk->nbytes;
  250. __serpent_encrypt(ctx, keystream, ctrblk);
  251. crypto_xor(keystream, src, nbytes);
  252. memcpy(dst, keystream, nbytes);
  253. crypto_inc(ctrblk, SERPENT_BLOCK_SIZE);
  254. }
  255. static unsigned int __ctr_crypt(struct blkcipher_desc *desc,
  256. struct blkcipher_walk *walk)
  257. {
  258. struct serpent_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  259. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  260. unsigned int nbytes = walk->nbytes;
  261. u128 *src = (u128 *)walk->src.virt.addr;
  262. u128 *dst = (u128 *)walk->dst.virt.addr;
  263. u128 ctrblk;
  264. be128 ctrblocks[SERPENT_PARALLEL_BLOCKS];
  265. int i;
  266. be128_to_u128(&ctrblk, (be128 *)walk->iv);
  267. /* Process multi-block batch */
  268. if (nbytes >= bsize * SERPENT_PARALLEL_BLOCKS) {
  269. do {
  270. /* create ctrblks for parallel encrypt */
  271. for (i = 0; i < SERPENT_PARALLEL_BLOCKS; i++) {
  272. if (dst != src)
  273. dst[i] = src[i];
  274. u128_to_be128(&ctrblocks[i], &ctrblk);
  275. u128_inc(&ctrblk);
  276. }
  277. serpent_enc_blk_xway_xor(ctx, (u8 *)dst,
  278. (u8 *)ctrblocks);
  279. src += SERPENT_PARALLEL_BLOCKS;
  280. dst += SERPENT_PARALLEL_BLOCKS;
  281. nbytes -= bsize * SERPENT_PARALLEL_BLOCKS;
  282. } while (nbytes >= bsize * SERPENT_PARALLEL_BLOCKS);
  283. if (nbytes < bsize)
  284. goto done;
  285. }
  286. /* Handle leftovers */
  287. do {
  288. if (dst != src)
  289. *dst = *src;
  290. u128_to_be128(&ctrblocks[0], &ctrblk);
  291. u128_inc(&ctrblk);
  292. __serpent_encrypt(ctx, (u8 *)ctrblocks, (u8 *)ctrblocks);
  293. u128_xor(dst, dst, (u128 *)ctrblocks);
  294. src += 1;
  295. dst += 1;
  296. nbytes -= bsize;
  297. } while (nbytes >= bsize);
  298. done:
  299. u128_to_be128((be128 *)walk->iv, &ctrblk);
  300. return nbytes;
  301. }
  302. static int ctr_crypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  303. struct scatterlist *src, unsigned int nbytes)
  304. {
  305. bool fpu_enabled = false;
  306. struct blkcipher_walk walk;
  307. int err;
  308. blkcipher_walk_init(&walk, dst, src, nbytes);
  309. err = blkcipher_walk_virt_block(desc, &walk, SERPENT_BLOCK_SIZE);
  310. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  311. while ((nbytes = walk.nbytes) >= SERPENT_BLOCK_SIZE) {
  312. fpu_enabled = serpent_fpu_begin(fpu_enabled, nbytes);
  313. nbytes = __ctr_crypt(desc, &walk);
  314. err = blkcipher_walk_done(desc, &walk, nbytes);
  315. }
  316. serpent_fpu_end(fpu_enabled);
  317. if (walk.nbytes) {
  318. ctr_crypt_final(desc, &walk);
  319. err = blkcipher_walk_done(desc, &walk, 0);
  320. }
  321. return err;
  322. }
  323. struct crypt_priv {
  324. struct serpent_ctx *ctx;
  325. bool fpu_enabled;
  326. };
  327. static void encrypt_callback(void *priv, u8 *srcdst, unsigned int nbytes)
  328. {
  329. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  330. struct crypt_priv *ctx = priv;
  331. int i;
  332. ctx->fpu_enabled = serpent_fpu_begin(ctx->fpu_enabled, nbytes);
  333. if (nbytes == bsize * SERPENT_PARALLEL_BLOCKS) {
  334. serpent_enc_blk_xway(ctx->ctx, srcdst, srcdst);
  335. return;
  336. }
  337. for (i = 0; i < nbytes / bsize; i++, srcdst += bsize)
  338. __serpent_encrypt(ctx->ctx, srcdst, srcdst);
  339. }
  340. static void decrypt_callback(void *priv, u8 *srcdst, unsigned int nbytes)
  341. {
  342. const unsigned int bsize = SERPENT_BLOCK_SIZE;
  343. struct crypt_priv *ctx = priv;
  344. int i;
  345. ctx->fpu_enabled = serpent_fpu_begin(ctx->fpu_enabled, nbytes);
  346. if (nbytes == bsize * SERPENT_PARALLEL_BLOCKS) {
  347. serpent_dec_blk_xway(ctx->ctx, srcdst, srcdst);
  348. return;
  349. }
  350. for (i = 0; i < nbytes / bsize; i++, srcdst += bsize)
  351. __serpent_decrypt(ctx->ctx, srcdst, srcdst);
  352. }
  353. struct serpent_lrw_ctx {
  354. struct lrw_table_ctx lrw_table;
  355. struct serpent_ctx serpent_ctx;
  356. };
  357. static int lrw_serpent_setkey(struct crypto_tfm *tfm, const u8 *key,
  358. unsigned int keylen)
  359. {
  360. struct serpent_lrw_ctx *ctx = crypto_tfm_ctx(tfm);
  361. int err;
  362. err = __serpent_setkey(&ctx->serpent_ctx, key, keylen -
  363. SERPENT_BLOCK_SIZE);
  364. if (err)
  365. return err;
  366. return lrw_init_table(&ctx->lrw_table, key + keylen -
  367. SERPENT_BLOCK_SIZE);
  368. }
  369. static int lrw_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  370. struct scatterlist *src, unsigned int nbytes)
  371. {
  372. struct serpent_lrw_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  373. be128 buf[SERPENT_PARALLEL_BLOCKS];
  374. struct crypt_priv crypt_ctx = {
  375. .ctx = &ctx->serpent_ctx,
  376. .fpu_enabled = false,
  377. };
  378. struct lrw_crypt_req req = {
  379. .tbuf = buf,
  380. .tbuflen = sizeof(buf),
  381. .table_ctx = &ctx->lrw_table,
  382. .crypt_ctx = &crypt_ctx,
  383. .crypt_fn = encrypt_callback,
  384. };
  385. int ret;
  386. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  387. ret = lrw_crypt(desc, dst, src, nbytes, &req);
  388. serpent_fpu_end(crypt_ctx.fpu_enabled);
  389. return ret;
  390. }
  391. static int lrw_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  392. struct scatterlist *src, unsigned int nbytes)
  393. {
  394. struct serpent_lrw_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  395. be128 buf[SERPENT_PARALLEL_BLOCKS];
  396. struct crypt_priv crypt_ctx = {
  397. .ctx = &ctx->serpent_ctx,
  398. .fpu_enabled = false,
  399. };
  400. struct lrw_crypt_req req = {
  401. .tbuf = buf,
  402. .tbuflen = sizeof(buf),
  403. .table_ctx = &ctx->lrw_table,
  404. .crypt_ctx = &crypt_ctx,
  405. .crypt_fn = decrypt_callback,
  406. };
  407. int ret;
  408. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  409. ret = lrw_crypt(desc, dst, src, nbytes, &req);
  410. serpent_fpu_end(crypt_ctx.fpu_enabled);
  411. return ret;
  412. }
  413. static void lrw_exit_tfm(struct crypto_tfm *tfm)
  414. {
  415. struct serpent_lrw_ctx *ctx = crypto_tfm_ctx(tfm);
  416. lrw_free_table(&ctx->lrw_table);
  417. }
  418. struct serpent_xts_ctx {
  419. struct serpent_ctx tweak_ctx;
  420. struct serpent_ctx crypt_ctx;
  421. };
  422. static int xts_serpent_setkey(struct crypto_tfm *tfm, const u8 *key,
  423. unsigned int keylen)
  424. {
  425. struct serpent_xts_ctx *ctx = crypto_tfm_ctx(tfm);
  426. u32 *flags = &tfm->crt_flags;
  427. int err;
  428. /* key consists of keys of equal size concatenated, therefore
  429. * the length must be even
  430. */
  431. if (keylen % 2) {
  432. *flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
  433. return -EINVAL;
  434. }
  435. /* first half of xts-key is for crypt */
  436. err = __serpent_setkey(&ctx->crypt_ctx, key, keylen / 2);
  437. if (err)
  438. return err;
  439. /* second half of xts-key is for tweak */
  440. return __serpent_setkey(&ctx->tweak_ctx, key + keylen / 2, keylen / 2);
  441. }
  442. static int xts_encrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  443. struct scatterlist *src, unsigned int nbytes)
  444. {
  445. struct serpent_xts_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  446. be128 buf[SERPENT_PARALLEL_BLOCKS];
  447. struct crypt_priv crypt_ctx = {
  448. .ctx = &ctx->crypt_ctx,
  449. .fpu_enabled = false,
  450. };
  451. struct xts_crypt_req req = {
  452. .tbuf = buf,
  453. .tbuflen = sizeof(buf),
  454. .tweak_ctx = &ctx->tweak_ctx,
  455. .tweak_fn = XTS_TWEAK_CAST(__serpent_encrypt),
  456. .crypt_ctx = &crypt_ctx,
  457. .crypt_fn = encrypt_callback,
  458. };
  459. int ret;
  460. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  461. ret = xts_crypt(desc, dst, src, nbytes, &req);
  462. serpent_fpu_end(crypt_ctx.fpu_enabled);
  463. return ret;
  464. }
  465. static int xts_decrypt(struct blkcipher_desc *desc, struct scatterlist *dst,
  466. struct scatterlist *src, unsigned int nbytes)
  467. {
  468. struct serpent_xts_ctx *ctx = crypto_blkcipher_ctx(desc->tfm);
  469. be128 buf[SERPENT_PARALLEL_BLOCKS];
  470. struct crypt_priv crypt_ctx = {
  471. .ctx = &ctx->crypt_ctx,
  472. .fpu_enabled = false,
  473. };
  474. struct xts_crypt_req req = {
  475. .tbuf = buf,
  476. .tbuflen = sizeof(buf),
  477. .tweak_ctx = &ctx->tweak_ctx,
  478. .tweak_fn = XTS_TWEAK_CAST(__serpent_encrypt),
  479. .crypt_ctx = &crypt_ctx,
  480. .crypt_fn = decrypt_callback,
  481. };
  482. int ret;
  483. desc->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP;
  484. ret = xts_crypt(desc, dst, src, nbytes, &req);
  485. serpent_fpu_end(crypt_ctx.fpu_enabled);
  486. return ret;
  487. }
  488. static struct crypto_alg serpent_algs[10] = { {
  489. .cra_name = "__ecb-serpent-sse2",
  490. .cra_driver_name = "__driver-ecb-serpent-sse2",
  491. .cra_priority = 0,
  492. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
  493. .cra_blocksize = SERPENT_BLOCK_SIZE,
  494. .cra_ctxsize = sizeof(struct serpent_ctx),
  495. .cra_alignmask = 0,
  496. .cra_type = &crypto_blkcipher_type,
  497. .cra_module = THIS_MODULE,
  498. .cra_list = LIST_HEAD_INIT(serpent_algs[0].cra_list),
  499. .cra_u = {
  500. .blkcipher = {
  501. .min_keysize = SERPENT_MIN_KEY_SIZE,
  502. .max_keysize = SERPENT_MAX_KEY_SIZE,
  503. .setkey = serpent_setkey,
  504. .encrypt = ecb_encrypt,
  505. .decrypt = ecb_decrypt,
  506. },
  507. },
  508. }, {
  509. .cra_name = "__cbc-serpent-sse2",
  510. .cra_driver_name = "__driver-cbc-serpent-sse2",
  511. .cra_priority = 0,
  512. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
  513. .cra_blocksize = SERPENT_BLOCK_SIZE,
  514. .cra_ctxsize = sizeof(struct serpent_ctx),
  515. .cra_alignmask = 0,
  516. .cra_type = &crypto_blkcipher_type,
  517. .cra_module = THIS_MODULE,
  518. .cra_list = LIST_HEAD_INIT(serpent_algs[1].cra_list),
  519. .cra_u = {
  520. .blkcipher = {
  521. .min_keysize = SERPENT_MIN_KEY_SIZE,
  522. .max_keysize = SERPENT_MAX_KEY_SIZE,
  523. .setkey = serpent_setkey,
  524. .encrypt = cbc_encrypt,
  525. .decrypt = cbc_decrypt,
  526. },
  527. },
  528. }, {
  529. .cra_name = "__ctr-serpent-sse2",
  530. .cra_driver_name = "__driver-ctr-serpent-sse2",
  531. .cra_priority = 0,
  532. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
  533. .cra_blocksize = 1,
  534. .cra_ctxsize = sizeof(struct serpent_ctx),
  535. .cra_alignmask = 0,
  536. .cra_type = &crypto_blkcipher_type,
  537. .cra_module = THIS_MODULE,
  538. .cra_list = LIST_HEAD_INIT(serpent_algs[2].cra_list),
  539. .cra_u = {
  540. .blkcipher = {
  541. .min_keysize = SERPENT_MIN_KEY_SIZE,
  542. .max_keysize = SERPENT_MAX_KEY_SIZE,
  543. .ivsize = SERPENT_BLOCK_SIZE,
  544. .setkey = serpent_setkey,
  545. .encrypt = ctr_crypt,
  546. .decrypt = ctr_crypt,
  547. },
  548. },
  549. }, {
  550. .cra_name = "__lrw-serpent-sse2",
  551. .cra_driver_name = "__driver-lrw-serpent-sse2",
  552. .cra_priority = 0,
  553. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
  554. .cra_blocksize = SERPENT_BLOCK_SIZE,
  555. .cra_ctxsize = sizeof(struct serpent_lrw_ctx),
  556. .cra_alignmask = 0,
  557. .cra_type = &crypto_blkcipher_type,
  558. .cra_module = THIS_MODULE,
  559. .cra_list = LIST_HEAD_INIT(serpent_algs[3].cra_list),
  560. .cra_exit = lrw_exit_tfm,
  561. .cra_u = {
  562. .blkcipher = {
  563. .min_keysize = SERPENT_MIN_KEY_SIZE +
  564. SERPENT_BLOCK_SIZE,
  565. .max_keysize = SERPENT_MAX_KEY_SIZE +
  566. SERPENT_BLOCK_SIZE,
  567. .ivsize = SERPENT_BLOCK_SIZE,
  568. .setkey = lrw_serpent_setkey,
  569. .encrypt = lrw_encrypt,
  570. .decrypt = lrw_decrypt,
  571. },
  572. },
  573. }, {
  574. .cra_name = "__xts-serpent-sse2",
  575. .cra_driver_name = "__driver-xts-serpent-sse2",
  576. .cra_priority = 0,
  577. .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER,
  578. .cra_blocksize = SERPENT_BLOCK_SIZE,
  579. .cra_ctxsize = sizeof(struct serpent_xts_ctx),
  580. .cra_alignmask = 0,
  581. .cra_type = &crypto_blkcipher_type,
  582. .cra_module = THIS_MODULE,
  583. .cra_list = LIST_HEAD_INIT(serpent_algs[4].cra_list),
  584. .cra_u = {
  585. .blkcipher = {
  586. .min_keysize = SERPENT_MIN_KEY_SIZE * 2,
  587. .max_keysize = SERPENT_MAX_KEY_SIZE * 2,
  588. .ivsize = SERPENT_BLOCK_SIZE,
  589. .setkey = xts_serpent_setkey,
  590. .encrypt = xts_encrypt,
  591. .decrypt = xts_decrypt,
  592. },
  593. },
  594. }, {
  595. .cra_name = "ecb(serpent)",
  596. .cra_driver_name = "ecb-serpent-sse2",
  597. .cra_priority = 400,
  598. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  599. .cra_blocksize = SERPENT_BLOCK_SIZE,
  600. .cra_ctxsize = sizeof(struct async_helper_ctx),
  601. .cra_alignmask = 0,
  602. .cra_type = &crypto_ablkcipher_type,
  603. .cra_module = THIS_MODULE,
  604. .cra_list = LIST_HEAD_INIT(serpent_algs[5].cra_list),
  605. .cra_init = ablk_init,
  606. .cra_exit = ablk_exit,
  607. .cra_u = {
  608. .ablkcipher = {
  609. .min_keysize = SERPENT_MIN_KEY_SIZE,
  610. .max_keysize = SERPENT_MAX_KEY_SIZE,
  611. .setkey = ablk_set_key,
  612. .encrypt = ablk_encrypt,
  613. .decrypt = ablk_decrypt,
  614. },
  615. },
  616. }, {
  617. .cra_name = "cbc(serpent)",
  618. .cra_driver_name = "cbc-serpent-sse2",
  619. .cra_priority = 400,
  620. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  621. .cra_blocksize = SERPENT_BLOCK_SIZE,
  622. .cra_ctxsize = sizeof(struct async_helper_ctx),
  623. .cra_alignmask = 0,
  624. .cra_type = &crypto_ablkcipher_type,
  625. .cra_module = THIS_MODULE,
  626. .cra_list = LIST_HEAD_INIT(serpent_algs[6].cra_list),
  627. .cra_init = ablk_init,
  628. .cra_exit = ablk_exit,
  629. .cra_u = {
  630. .ablkcipher = {
  631. .min_keysize = SERPENT_MIN_KEY_SIZE,
  632. .max_keysize = SERPENT_MAX_KEY_SIZE,
  633. .ivsize = SERPENT_BLOCK_SIZE,
  634. .setkey = ablk_set_key,
  635. .encrypt = __ablk_encrypt,
  636. .decrypt = ablk_decrypt,
  637. },
  638. },
  639. }, {
  640. .cra_name = "ctr(serpent)",
  641. .cra_driver_name = "ctr-serpent-sse2",
  642. .cra_priority = 400,
  643. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  644. .cra_blocksize = 1,
  645. .cra_ctxsize = sizeof(struct async_helper_ctx),
  646. .cra_alignmask = 0,
  647. .cra_type = &crypto_ablkcipher_type,
  648. .cra_module = THIS_MODULE,
  649. .cra_list = LIST_HEAD_INIT(serpent_algs[7].cra_list),
  650. .cra_init = ablk_init,
  651. .cra_exit = ablk_exit,
  652. .cra_u = {
  653. .ablkcipher = {
  654. .min_keysize = SERPENT_MIN_KEY_SIZE,
  655. .max_keysize = SERPENT_MAX_KEY_SIZE,
  656. .ivsize = SERPENT_BLOCK_SIZE,
  657. .setkey = ablk_set_key,
  658. .encrypt = ablk_encrypt,
  659. .decrypt = ablk_encrypt,
  660. .geniv = "chainiv",
  661. },
  662. },
  663. }, {
  664. .cra_name = "lrw(serpent)",
  665. .cra_driver_name = "lrw-serpent-sse2",
  666. .cra_priority = 400,
  667. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  668. .cra_blocksize = SERPENT_BLOCK_SIZE,
  669. .cra_ctxsize = sizeof(struct async_helper_ctx),
  670. .cra_alignmask = 0,
  671. .cra_type = &crypto_ablkcipher_type,
  672. .cra_module = THIS_MODULE,
  673. .cra_list = LIST_HEAD_INIT(serpent_algs[8].cra_list),
  674. .cra_init = ablk_init,
  675. .cra_exit = ablk_exit,
  676. .cra_u = {
  677. .ablkcipher = {
  678. .min_keysize = SERPENT_MIN_KEY_SIZE +
  679. SERPENT_BLOCK_SIZE,
  680. .max_keysize = SERPENT_MAX_KEY_SIZE +
  681. SERPENT_BLOCK_SIZE,
  682. .ivsize = SERPENT_BLOCK_SIZE,
  683. .setkey = ablk_set_key,
  684. .encrypt = ablk_encrypt,
  685. .decrypt = ablk_decrypt,
  686. },
  687. },
  688. }, {
  689. .cra_name = "xts(serpent)",
  690. .cra_driver_name = "xts-serpent-sse2",
  691. .cra_priority = 400,
  692. .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC,
  693. .cra_blocksize = SERPENT_BLOCK_SIZE,
  694. .cra_ctxsize = sizeof(struct async_helper_ctx),
  695. .cra_alignmask = 0,
  696. .cra_type = &crypto_ablkcipher_type,
  697. .cra_module = THIS_MODULE,
  698. .cra_list = LIST_HEAD_INIT(serpent_algs[9].cra_list),
  699. .cra_init = ablk_init,
  700. .cra_exit = ablk_exit,
  701. .cra_u = {
  702. .ablkcipher = {
  703. .min_keysize = SERPENT_MIN_KEY_SIZE * 2,
  704. .max_keysize = SERPENT_MAX_KEY_SIZE * 2,
  705. .ivsize = SERPENT_BLOCK_SIZE,
  706. .setkey = ablk_set_key,
  707. .encrypt = ablk_encrypt,
  708. .decrypt = ablk_decrypt,
  709. },
  710. },
  711. } };
  712. static int __init serpent_sse2_init(void)
  713. {
  714. if (!cpu_has_xmm2) {
  715. printk(KERN_INFO "SSE2 instructions are not detected.\n");
  716. return -ENODEV;
  717. }
  718. return crypto_register_algs(serpent_algs, ARRAY_SIZE(serpent_algs));
  719. }
  720. static void __exit serpent_sse2_exit(void)
  721. {
  722. crypto_unregister_algs(serpent_algs, ARRAY_SIZE(serpent_algs));
  723. }
  724. module_init(serpent_sse2_init);
  725. module_exit(serpent_sse2_exit);
  726. MODULE_DESCRIPTION("Serpent Cipher Algorithm, SSE2 optimized");
  727. MODULE_LICENSE("GPL");
  728. MODULE_ALIAS("serpent");