7c8c3e591c
SVN-Revision: 32672
1323 lines
36 KiB
C
1323 lines
36 KiB
C
/*
|
|
* An OCF module that uses the linux kernel cryptoapi, based on the
|
|
* original cryptosoft for BSD by Angelos D. Keromytis (angelos@cis.upenn.edu)
|
|
* but is mostly unrecognisable,
|
|
*
|
|
* Written by David McCullough <david_mccullough@mcafee.com>
|
|
* Copyright (C) 2004-2011 David McCullough
|
|
* Copyright (C) 2004-2005 Intel Corporation.
|
|
*
|
|
* LICENSE TERMS
|
|
*
|
|
* The free distribution and use of this software in both source and binary
|
|
* form is allowed (with or without changes) provided that:
|
|
*
|
|
* 1. distributions of this source code include the above copyright
|
|
* notice, this list of conditions and the following disclaimer;
|
|
*
|
|
* 2. distributions in binary form include the above copyright
|
|
* notice, this list of conditions and the following disclaimer
|
|
* in the documentation and/or other associated materials;
|
|
*
|
|
* 3. the copyright holder's name is not used to endorse products
|
|
* built using this software without specific written permission.
|
|
*
|
|
* ALTERNATIVELY, provided that this notice is retained in full, this product
|
|
* may be distributed under the terms of the GNU General Public License (GPL),
|
|
* in which case the provisions of the GPL apply INSTEAD OF those given above.
|
|
*
|
|
* DISCLAIMER
|
|
*
|
|
* This software is provided 'as is' with no explicit or implied warranties
|
|
* in respect of its properties, including, but not limited to, correctness
|
|
* and/or fitness for purpose.
|
|
* ---------------------------------------------------------------------------
|
|
*/
|
|
|
|
#include <linux/version.h>
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,38) && !defined(AUTOCONF_INCLUDED)
|
|
#include <linux/config.h>
|
|
#endif
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/list.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/crypto.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/skbuff.h>
|
|
#include <linux/random.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/spinlock.h>
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,10)
|
|
#include <linux/scatterlist.h>
|
|
#endif
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,29)
|
|
#include <crypto/hash.h>
|
|
#endif
|
|
|
|
#include <cryptodev.h>
|
|
#include <uio.h>
|
|
|
|
struct {
|
|
softc_device_decl sc_dev;
|
|
} swcr_softc;
|
|
|
|
#define offset_in_page(p) ((unsigned long)(p) & ~PAGE_MASK)
|
|
|
|
#define SW_TYPE_CIPHER 0x01
|
|
#define SW_TYPE_HMAC 0x02
|
|
#define SW_TYPE_HASH 0x04
|
|
#define SW_TYPE_COMP 0x08
|
|
#define SW_TYPE_BLKCIPHER 0x10
|
|
#define SW_TYPE_ALG_MASK 0x1f
|
|
|
|
#define SW_TYPE_ASYNC 0x8000
|
|
|
|
#define SW_TYPE_INUSE 0x10000000
|
|
|
|
/* We change some of the above if we have an async interface */
|
|
|
|
#define SW_TYPE_ALG_AMASK (SW_TYPE_ALG_MASK | SW_TYPE_ASYNC)
|
|
|
|
#define SW_TYPE_ABLKCIPHER (SW_TYPE_BLKCIPHER | SW_TYPE_ASYNC)
|
|
#define SW_TYPE_AHASH (SW_TYPE_HASH | SW_TYPE_ASYNC)
|
|
#define SW_TYPE_AHMAC (SW_TYPE_HMAC | SW_TYPE_ASYNC)
|
|
|
|
#define SCATTERLIST_MAX 16
|
|
|
|
struct swcr_data {
|
|
struct work_struct workq;
|
|
int sw_type;
|
|
int sw_alg;
|
|
struct crypto_tfm *sw_tfm;
|
|
spinlock_t sw_tfm_lock;
|
|
union {
|
|
struct {
|
|
char *sw_key;
|
|
int sw_klen;
|
|
int sw_mlen;
|
|
} hmac;
|
|
void *sw_comp_buf;
|
|
} u;
|
|
struct swcr_data *sw_next;
|
|
};
|
|
|
|
struct swcr_req {
|
|
struct swcr_data *sw_head;
|
|
struct swcr_data *sw;
|
|
struct cryptop *crp;
|
|
struct cryptodesc *crd;
|
|
struct scatterlist sg[SCATTERLIST_MAX];
|
|
unsigned char iv[EALG_MAX_BLOCK_LEN];
|
|
char result[HASH_MAX_LEN];
|
|
void *crypto_req;
|
|
};
|
|
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
|
|
static kmem_cache_t *swcr_req_cache;
|
|
#else
|
|
static struct kmem_cache *swcr_req_cache;
|
|
#endif
|
|
|
|
#ifndef CRYPTO_TFM_MODE_CBC
|
|
/*
|
|
* As of linux-2.6.21 this is no longer defined, and presumably no longer
|
|
* needed to be passed into the crypto core code.
|
|
*/
|
|
#define CRYPTO_TFM_MODE_CBC 0
|
|
#define CRYPTO_TFM_MODE_ECB 0
|
|
#endif
|
|
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,19)
|
|
/*
|
|
* Linux 2.6.19 introduced a new Crypto API, setup macro's to convert new
|
|
* API into old API.
|
|
*/
|
|
|
|
/* Symmetric/Block Cipher */
|
|
struct blkcipher_desc
|
|
{
|
|
struct crypto_tfm *tfm;
|
|
void *info;
|
|
};
|
|
#define ecb(X) #X , CRYPTO_TFM_MODE_ECB
|
|
#define cbc(X) #X , CRYPTO_TFM_MODE_CBC
|
|
#define crypto_has_blkcipher(X, Y, Z) crypto_alg_available(X, 0)
|
|
#define crypto_blkcipher_cast(X) X
|
|
#define crypto_blkcipher_tfm(X) X
|
|
#define crypto_alloc_blkcipher(X, Y, Z) crypto_alloc_tfm(X, mode)
|
|
#define crypto_blkcipher_ivsize(X) crypto_tfm_alg_ivsize(X)
|
|
#define crypto_blkcipher_blocksize(X) crypto_tfm_alg_blocksize(X)
|
|
#define crypto_blkcipher_setkey(X, Y, Z) crypto_cipher_setkey(X, Y, Z)
|
|
#define crypto_blkcipher_encrypt_iv(W, X, Y, Z) \
|
|
crypto_cipher_encrypt_iv((W)->tfm, X, Y, Z, (u8 *)((W)->info))
|
|
#define crypto_blkcipher_decrypt_iv(W, X, Y, Z) \
|
|
crypto_cipher_decrypt_iv((W)->tfm, X, Y, Z, (u8 *)((W)->info))
|
|
#define crypto_blkcipher_set_flags(x, y) /* nop */
|
|
#define crypto_free_blkcipher(x) crypto_free_tfm(x)
|
|
#define crypto_free_comp crypto_free_tfm
|
|
#define crypto_free_hash crypto_free_tfm
|
|
|
|
/* Hash/HMAC/Digest */
|
|
struct hash_desc
|
|
{
|
|
struct crypto_tfm *tfm;
|
|
};
|
|
#define hmac(X) #X , 0
|
|
#define crypto_has_hash(X, Y, Z) crypto_alg_available(X, 0)
|
|
#define crypto_hash_cast(X) X
|
|
#define crypto_hash_tfm(X) X
|
|
#define crypto_alloc_hash(X, Y, Z) crypto_alloc_tfm(X, mode)
|
|
#define crypto_hash_digestsize(X) crypto_tfm_alg_digestsize(X)
|
|
#define crypto_hash_digest(W, X, Y, Z) \
|
|
crypto_digest_digest((W)->tfm, X, sg_num, Z)
|
|
|
|
/* Asymmetric Cipher */
|
|
#define crypto_has_cipher(X, Y, Z) crypto_alg_available(X, 0)
|
|
|
|
/* Compression */
|
|
#define crypto_has_comp(X, Y, Z) crypto_alg_available(X, 0)
|
|
#define crypto_comp_tfm(X) X
|
|
#define crypto_comp_cast(X) X
|
|
#define crypto_alloc_comp(X, Y, Z) crypto_alloc_tfm(X, mode)
|
|
#define plain(X) #X , 0
|
|
#else
|
|
#define ecb(X) "ecb(" #X ")" , 0
|
|
#define cbc(X) "cbc(" #X ")" , 0
|
|
#define hmac(X) "hmac(" #X ")" , 0
|
|
#define plain(X) #X , 0
|
|
#endif /* if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,19) */
|
|
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,22)
|
|
/* no ablkcipher in older kernels */
|
|
#define crypto_alloc_ablkcipher(a,b,c) (NULL)
|
|
#define crypto_ablkcipher_tfm(x) ((struct crypto_tfm *)(x))
|
|
#define crypto_ablkcipher_set_flags(a, b) /* nop */
|
|
#define crypto_ablkcipher_setkey(x, y, z) (-EINVAL)
|
|
#define crypto_has_ablkcipher(a,b,c) (0)
|
|
#else
|
|
#define HAVE_ABLKCIPHER
|
|
#endif
|
|
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,32)
|
|
/* no ahash in older kernels */
|
|
#define crypto_ahash_tfm(x) ((struct crypto_tfm *)(x))
|
|
#define crypto_alloc_ahash(a,b,c) (NULL)
|
|
#define crypto_ahash_digestsize(x) 0
|
|
#else
|
|
#define HAVE_AHASH
|
|
#endif
|
|
|
|
struct crypto_details {
|
|
char *alg_name;
|
|
int mode;
|
|
int sw_type;
|
|
};
|
|
|
|
static struct crypto_details crypto_details[] = {
|
|
[CRYPTO_DES_CBC] = { cbc(des), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_3DES_CBC] = { cbc(des3_ede), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_BLF_CBC] = { cbc(blowfish), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_CAST_CBC] = { cbc(cast5), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_SKIPJACK_CBC] = { cbc(skipjack), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_MD5_HMAC] = { hmac(md5), SW_TYPE_HMAC, },
|
|
[CRYPTO_SHA1_HMAC] = { hmac(sha1), SW_TYPE_HMAC, },
|
|
[CRYPTO_RIPEMD160_HMAC] = { hmac(ripemd160), SW_TYPE_HMAC, },
|
|
[CRYPTO_MD5_KPDK] = { plain(md5-kpdk), SW_TYPE_HASH, },
|
|
[CRYPTO_SHA1_KPDK] = { plain(sha1-kpdk), SW_TYPE_HASH, },
|
|
[CRYPTO_AES_CBC] = { cbc(aes), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_ARC4] = { ecb(arc4), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_MD5] = { plain(md5), SW_TYPE_HASH, },
|
|
[CRYPTO_SHA1] = { plain(sha1), SW_TYPE_HASH, },
|
|
[CRYPTO_NULL_HMAC] = { hmac(digest_null), SW_TYPE_HMAC, },
|
|
[CRYPTO_NULL_CBC] = { cbc(cipher_null), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_DEFLATE_COMP] = { plain(deflate), SW_TYPE_COMP, },
|
|
[CRYPTO_SHA2_256_HMAC] = { hmac(sha256), SW_TYPE_HMAC, },
|
|
[CRYPTO_SHA2_384_HMAC] = { hmac(sha384), SW_TYPE_HMAC, },
|
|
[CRYPTO_SHA2_512_HMAC] = { hmac(sha512), SW_TYPE_HMAC, },
|
|
[CRYPTO_CAMELLIA_CBC] = { cbc(camellia), SW_TYPE_BLKCIPHER, },
|
|
[CRYPTO_SHA2_256] = { plain(sha256), SW_TYPE_HASH, },
|
|
[CRYPTO_SHA2_384] = { plain(sha384), SW_TYPE_HASH, },
|
|
[CRYPTO_SHA2_512] = { plain(sha512), SW_TYPE_HASH, },
|
|
[CRYPTO_RIPEMD160] = { plain(ripemd160), SW_TYPE_HASH, },
|
|
};
|
|
|
|
int32_t swcr_id = -1;
|
|
module_param(swcr_id, int, 0444);
|
|
MODULE_PARM_DESC(swcr_id, "Read-Only OCF ID for cryptosoft driver");
|
|
|
|
int swcr_fail_if_compression_grows = 1;
|
|
module_param(swcr_fail_if_compression_grows, int, 0644);
|
|
MODULE_PARM_DESC(swcr_fail_if_compression_grows,
|
|
"Treat compression that results in more data as a failure");
|
|
|
|
int swcr_no_ahash = 0;
|
|
module_param(swcr_no_ahash, int, 0644);
|
|
MODULE_PARM_DESC(swcr_no_ahash,
|
|
"Do not use async hash/hmac even if available");
|
|
|
|
int swcr_no_ablk = 0;
|
|
module_param(swcr_no_ablk, int, 0644);
|
|
MODULE_PARM_DESC(swcr_no_ablk,
|
|
"Do not use async blk ciphers even if available");
|
|
|
|
static struct swcr_data **swcr_sessions = NULL;
|
|
static u_int32_t swcr_sesnum = 0;
|
|
|
|
static int swcr_process(device_t, struct cryptop *, int);
|
|
static int swcr_newsession(device_t, u_int32_t *, struct cryptoini *);
|
|
static int swcr_freesession(device_t, u_int64_t);
|
|
|
|
static device_method_t swcr_methods = {
|
|
/* crypto device methods */
|
|
DEVMETHOD(cryptodev_newsession, swcr_newsession),
|
|
DEVMETHOD(cryptodev_freesession,swcr_freesession),
|
|
DEVMETHOD(cryptodev_process, swcr_process),
|
|
};
|
|
|
|
#define debug swcr_debug
|
|
int swcr_debug = 0;
|
|
module_param(swcr_debug, int, 0644);
|
|
MODULE_PARM_DESC(swcr_debug, "Enable debug");
|
|
|
|
static void swcr_process_req(struct swcr_req *req);
|
|
|
|
/*
|
|
* somethings just need to be run with user context no matter whether
|
|
* the kernel compression libs use vmalloc/vfree for example.
|
|
*/
|
|
|
|
typedef struct {
|
|
struct work_struct wq;
|
|
void (*func)(void *arg);
|
|
void *arg;
|
|
} execute_later_t;
|
|
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,20)
|
|
static void
|
|
doing_it_now(struct work_struct *wq)
|
|
{
|
|
execute_later_t *w = container_of(wq, execute_later_t, wq);
|
|
(w->func)(w->arg);
|
|
kfree(w);
|
|
}
|
|
#else
|
|
static void
|
|
doing_it_now(void *arg)
|
|
{
|
|
execute_later_t *w = (execute_later_t *) arg;
|
|
(w->func)(w->arg);
|
|
kfree(w);
|
|
}
|
|
#endif
|
|
|
|
static void
|
|
execute_later(void (fn)(void *), void *arg)
|
|
{
|
|
execute_later_t *w;
|
|
|
|
w = (execute_later_t *) kmalloc(sizeof(execute_later_t), SLAB_ATOMIC);
|
|
if (w) {
|
|
memset(w, '\0', sizeof(w));
|
|
w->func = fn;
|
|
w->arg = arg;
|
|
#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,20)
|
|
INIT_WORK(&w->wq, doing_it_now);
|
|
#else
|
|
INIT_WORK(&w->wq, doing_it_now, w);
|
|
#endif
|
|
schedule_work(&w->wq);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Generate a new software session.
|
|
*/
|
|
static int
|
|
swcr_newsession(device_t dev, u_int32_t *sid, struct cryptoini *cri)
|
|
{
|
|
struct swcr_data **swd;
|
|
u_int32_t i;
|
|
int error;
|
|
char *algo;
|
|
int mode;
|
|
|
|
dprintk("%s()\n", __FUNCTION__);
|
|
if (sid == NULL || cri == NULL) {
|
|
dprintk("%s,%d - EINVAL\n", __FILE__, __LINE__);
|
|
return EINVAL;
|
|
}
|
|
|
|
if (swcr_sessions) {
|
|
for (i = 1; i < swcr_sesnum; i++)
|
|
if (swcr_sessions[i] == NULL)
|
|
break;
|
|
} else
|
|
i = 1; /* NB: to silence compiler warning */
|
|
|
|
if (swcr_sessions == NULL || i == swcr_sesnum) {
|
|
if (swcr_sessions == NULL) {
|
|
i = 1; /* We leave swcr_sessions[0] empty */
|
|
swcr_sesnum = CRYPTO_SW_SESSIONS;
|
|
} else
|
|
swcr_sesnum *= 2;
|
|
|
|
swd = kmalloc(swcr_sesnum * sizeof(struct swcr_data *), SLAB_ATOMIC);
|
|
if (swd == NULL) {
|
|
/* Reset session number */
|
|
if (swcr_sesnum == CRYPTO_SW_SESSIONS)
|
|
swcr_sesnum = 0;
|
|
else
|
|
swcr_sesnum /= 2;
|
|
dprintk("%s,%d: ENOBUFS\n", __FILE__, __LINE__);
|
|
return ENOBUFS;
|
|
}
|
|
memset(swd, 0, swcr_sesnum * sizeof(struct swcr_data *));
|
|
|
|
/* Copy existing sessions */
|
|
if (swcr_sessions) {
|
|
memcpy(swd, swcr_sessions,
|
|
(swcr_sesnum / 2) * sizeof(struct swcr_data *));
|
|
kfree(swcr_sessions);
|
|
}
|
|
|
|
swcr_sessions = swd;
|
|
}
|
|
|
|
swd = &swcr_sessions[i];
|
|
*sid = i;
|
|
|
|
while (cri) {
|
|
*swd = (struct swcr_data *) kmalloc(sizeof(struct swcr_data),
|
|
SLAB_ATOMIC);
|
|
if (*swd == NULL) {
|
|
swcr_freesession(NULL, i);
|
|
dprintk("%s,%d: ENOBUFS\n", __FILE__, __LINE__);
|
|
return ENOBUFS;
|
|
}
|
|
memset(*swd, 0, sizeof(struct swcr_data));
|
|
|
|
if (cri->cri_alg < 0 ||
|
|
cri->cri_alg>=sizeof(crypto_details)/sizeof(crypto_details[0])){
|
|
printk("cryptosoft: Unknown algorithm 0x%x\n", cri->cri_alg);
|
|
swcr_freesession(NULL, i);
|
|
return EINVAL;
|
|
}
|
|
|
|
algo = crypto_details[cri->cri_alg].alg_name;
|
|
if (!algo || !*algo) {
|
|
printk("cryptosoft: Unsupported algorithm 0x%x\n", cri->cri_alg);
|
|
swcr_freesession(NULL, i);
|
|
return EINVAL;
|
|
}
|
|
|
|
mode = crypto_details[cri->cri_alg].mode;
|
|
(*swd)->sw_type = crypto_details[cri->cri_alg].sw_type;
|
|
(*swd)->sw_alg = cri->cri_alg;
|
|
|
|
spin_lock_init(&(*swd)->sw_tfm_lock);
|
|
|
|
/* Algorithm specific configuration */
|
|
switch (cri->cri_alg) {
|
|
case CRYPTO_NULL_CBC:
|
|
cri->cri_klen = 0; /* make it work with crypto API */
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if ((*swd)->sw_type & SW_TYPE_BLKCIPHER) {
|
|
dprintk("%s crypto_alloc_*blkcipher(%s, 0x%x)\n", __FUNCTION__,
|
|
algo, mode);
|
|
|
|
/* try async first */
|
|
(*swd)->sw_tfm = swcr_no_ablk ? NULL :
|
|
crypto_ablkcipher_tfm(crypto_alloc_ablkcipher(algo, 0, 0));
|
|
if ((*swd)->sw_tfm && !IS_ERR((*swd)->sw_tfm)) {
|
|
dprintk("%s %s cipher is async\n", __FUNCTION__, algo);
|
|
(*swd)->sw_type |= SW_TYPE_ASYNC;
|
|
} else {
|
|
(*swd)->sw_tfm = crypto_blkcipher_tfm(
|
|
crypto_alloc_blkcipher(algo, 0, CRYPTO_ALG_ASYNC));
|
|
if ((*swd)->sw_tfm && !IS_ERR((*swd)->sw_tfm))
|
|
dprintk("%s %s cipher is sync\n", __FUNCTION__, algo);
|
|
}
|
|
if (!(*swd)->sw_tfm || IS_ERR((*swd)->sw_tfm)) {
|
|
int err;
|
|
dprintk("cryptosoft: crypto_alloc_blkcipher failed(%s, 0x%x)\n",
|
|
algo,mode);
|
|
err = IS_ERR((*swd)->sw_tfm) ? -(PTR_ERR((*swd)->sw_tfm)) : EINVAL;
|
|
(*swd)->sw_tfm = NULL; /* ensure NULL */
|
|
swcr_freesession(NULL, i);
|
|
return err;
|
|
}
|
|
|
|
if (debug) {
|
|
dprintk("%s key:cri->cri_klen=%d,(cri->cri_klen + 7)/8=%d",
|
|
__FUNCTION__, cri->cri_klen, (cri->cri_klen + 7) / 8);
|
|
for (i = 0; i < (cri->cri_klen + 7) / 8; i++)
|
|
dprintk("%s0x%x", (i % 8) ? " " : "\n ",
|
|
cri->cri_key[i] & 0xff);
|
|
dprintk("\n");
|
|
}
|
|
if ((*swd)->sw_type & SW_TYPE_ASYNC) {
|
|
/* OCF doesn't enforce keys */
|
|
crypto_ablkcipher_set_flags(
|
|
__crypto_ablkcipher_cast((*swd)->sw_tfm),
|
|
CRYPTO_TFM_REQ_WEAK_KEY);
|
|
error = crypto_ablkcipher_setkey(
|
|
__crypto_ablkcipher_cast((*swd)->sw_tfm),
|
|
cri->cri_key, (cri->cri_klen + 7) / 8);
|
|
} else {
|
|
/* OCF doesn't enforce keys */
|
|
crypto_blkcipher_set_flags(
|
|
crypto_blkcipher_cast((*swd)->sw_tfm),
|
|
CRYPTO_TFM_REQ_WEAK_KEY);
|
|
error = crypto_blkcipher_setkey(
|
|
crypto_blkcipher_cast((*swd)->sw_tfm),
|
|
cri->cri_key, (cri->cri_klen + 7) / 8);
|
|
}
|
|
if (error) {
|
|
printk("cryptosoft: setkey failed %d (crt_flags=0x%x)\n", error,
|
|
(*swd)->sw_tfm->crt_flags);
|
|
swcr_freesession(NULL, i);
|
|
return error;
|
|
}
|
|
} else if ((*swd)->sw_type & (SW_TYPE_HMAC | SW_TYPE_HASH)) {
|
|
dprintk("%s crypto_alloc_*hash(%s, 0x%x)\n", __FUNCTION__,
|
|
algo, mode);
|
|
|
|
/* try async first */
|
|
(*swd)->sw_tfm = swcr_no_ahash ? NULL :
|
|
crypto_ahash_tfm(crypto_alloc_ahash(algo, 0, 0));
|
|
if ((*swd)->sw_tfm) {
|
|
dprintk("%s %s hash is async\n", __FUNCTION__, algo);
|
|
(*swd)->sw_type |= SW_TYPE_ASYNC;
|
|
} else {
|
|
dprintk("%s %s hash is sync\n", __FUNCTION__, algo);
|
|
(*swd)->sw_tfm = crypto_hash_tfm(
|
|
crypto_alloc_hash(algo, 0, CRYPTO_ALG_ASYNC));
|
|
}
|
|
|
|
if (!(*swd)->sw_tfm) {
|
|
dprintk("cryptosoft: crypto_alloc_hash failed(%s,0x%x)\n",
|
|
algo, mode);
|
|
swcr_freesession(NULL, i);
|
|
return EINVAL;
|
|
}
|
|
|
|
(*swd)->u.hmac.sw_klen = (cri->cri_klen + 7) / 8;
|
|
(*swd)->u.hmac.sw_key = (char *)kmalloc((*swd)->u.hmac.sw_klen,
|
|
SLAB_ATOMIC);
|
|
if ((*swd)->u.hmac.sw_key == NULL) {
|
|
swcr_freesession(NULL, i);
|
|
dprintk("%s,%d: ENOBUFS\n", __FILE__, __LINE__);
|
|
return ENOBUFS;
|
|
}
|
|
memcpy((*swd)->u.hmac.sw_key, cri->cri_key, (*swd)->u.hmac.sw_klen);
|
|
if (cri->cri_mlen) {
|
|
(*swd)->u.hmac.sw_mlen = cri->cri_mlen;
|
|
} else if ((*swd)->sw_type & SW_TYPE_ASYNC) {
|
|
(*swd)->u.hmac.sw_mlen = crypto_ahash_digestsize(
|
|
__crypto_ahash_cast((*swd)->sw_tfm));
|
|
} else {
|
|
(*swd)->u.hmac.sw_mlen = crypto_hash_digestsize(
|
|
crypto_hash_cast((*swd)->sw_tfm));
|
|
}
|
|
} else if ((*swd)->sw_type & SW_TYPE_COMP) {
|
|
(*swd)->sw_tfm = crypto_comp_tfm(
|
|
crypto_alloc_comp(algo, 0, CRYPTO_ALG_ASYNC));
|
|
if (!(*swd)->sw_tfm) {
|
|
dprintk("cryptosoft: crypto_alloc_comp failed(%s,0x%x)\n",
|
|
algo, mode);
|
|
swcr_freesession(NULL, i);
|
|
return EINVAL;
|
|
}
|
|
(*swd)->u.sw_comp_buf = kmalloc(CRYPTO_MAX_DATA_LEN, SLAB_ATOMIC);
|
|
if ((*swd)->u.sw_comp_buf == NULL) {
|
|
swcr_freesession(NULL, i);
|
|
dprintk("%s,%d: ENOBUFS\n", __FILE__, __LINE__);
|
|
return ENOBUFS;
|
|
}
|
|
} else {
|
|
printk("cryptosoft: Unhandled sw_type %d\n", (*swd)->sw_type);
|
|
swcr_freesession(NULL, i);
|
|
return EINVAL;
|
|
}
|
|
|
|
cri = cri->cri_next;
|
|
swd = &((*swd)->sw_next);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Free a session.
|
|
*/
|
|
static int
|
|
swcr_freesession(device_t dev, u_int64_t tid)
|
|
{
|
|
struct swcr_data *swd;
|
|
u_int32_t sid = CRYPTO_SESID2LID(tid);
|
|
|
|
dprintk("%s()\n", __FUNCTION__);
|
|
if (sid > swcr_sesnum || swcr_sessions == NULL ||
|
|
swcr_sessions[sid] == NULL) {
|
|
dprintk("%s,%d: EINVAL\n", __FILE__, __LINE__);
|
|
return(EINVAL);
|
|
}
|
|
|
|
/* Silently accept and return */
|
|
if (sid == 0)
|
|
return(0);
|
|
|
|
while ((swd = swcr_sessions[sid]) != NULL) {
|
|
swcr_sessions[sid] = swd->sw_next;
|
|
if (swd->sw_tfm) {
|
|
switch (swd->sw_type & SW_TYPE_ALG_AMASK) {
|
|
#ifdef HAVE_AHASH
|
|
case SW_TYPE_AHMAC:
|
|
case SW_TYPE_AHASH:
|
|
crypto_free_ahash(__crypto_ahash_cast(swd->sw_tfm));
|
|
break;
|
|
#endif
|
|
#ifdef HAVE_ABLKCIPHER
|
|
case SW_TYPE_ABLKCIPHER:
|
|
crypto_free_ablkcipher(__crypto_ablkcipher_cast(swd->sw_tfm));
|
|
break;
|
|
#endif
|
|
case SW_TYPE_BLKCIPHER:
|
|
crypto_free_blkcipher(crypto_blkcipher_cast(swd->sw_tfm));
|
|
break;
|
|
case SW_TYPE_HMAC:
|
|
case SW_TYPE_HASH:
|
|
crypto_free_hash(crypto_hash_cast(swd->sw_tfm));
|
|
break;
|
|
case SW_TYPE_COMP:
|
|
if (in_interrupt())
|
|
execute_later((void (*)(void *))crypto_free_comp, (void *)crypto_comp_cast(swd->sw_tfm));
|
|
else
|
|
crypto_free_comp(crypto_comp_cast(swd->sw_tfm));
|
|
break;
|
|
default:
|
|
crypto_free_tfm(swd->sw_tfm);
|
|
break;
|
|
}
|
|
swd->sw_tfm = NULL;
|
|
}
|
|
if (swd->sw_type & SW_TYPE_COMP) {
|
|
if (swd->u.sw_comp_buf)
|
|
kfree(swd->u.sw_comp_buf);
|
|
} else {
|
|
if (swd->u.hmac.sw_key)
|
|
kfree(swd->u.hmac.sw_key);
|
|
}
|
|
kfree(swd);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void swcr_process_req_complete(struct swcr_req *req)
|
|
{
|
|
dprintk("%s()\n", __FUNCTION__);
|
|
|
|
if (req->sw->sw_type & SW_TYPE_INUSE) {
|
|
unsigned long flags;
|
|
spin_lock_irqsave(&req->sw->sw_tfm_lock, flags);
|
|
req->sw->sw_type &= ~SW_TYPE_INUSE;
|
|
spin_unlock_irqrestore(&req->sw->sw_tfm_lock, flags);
|
|
}
|
|
|
|
if (req->crp->crp_etype)
|
|
goto done;
|
|
|
|
switch (req->sw->sw_type & SW_TYPE_ALG_AMASK) {
|
|
#if defined(HAVE_AHASH)
|
|
case SW_TYPE_AHMAC:
|
|
case SW_TYPE_AHASH:
|
|
crypto_copyback(req->crp->crp_flags, req->crp->crp_buf,
|
|
req->crd->crd_inject, req->sw->u.hmac.sw_mlen, req->result);
|
|
ahash_request_free(req->crypto_req);
|
|
break;
|
|
#endif
|
|
#if defined(HAVE_ABLKCIPHER)
|
|
case SW_TYPE_ABLKCIPHER:
|
|
ablkcipher_request_free(req->crypto_req);
|
|
break;
|
|
#endif
|
|
case SW_TYPE_CIPHER:
|
|
case SW_TYPE_HMAC:
|
|
case SW_TYPE_HASH:
|
|
case SW_TYPE_COMP:
|
|
case SW_TYPE_BLKCIPHER:
|
|
break;
|
|
default:
|
|
req->crp->crp_etype = EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
req->crd = req->crd->crd_next;
|
|
if (req->crd) {
|
|
swcr_process_req(req);
|
|
return;
|
|
}
|
|
|
|
done:
|
|
dprintk("%s crypto_done %p\n", __FUNCTION__, req);
|
|
crypto_done(req->crp);
|
|
kmem_cache_free(swcr_req_cache, req);
|
|
}
|
|
|
|
#if defined(HAVE_ABLKCIPHER) || defined(HAVE_AHASH)
|
|
static void swcr_process_callback(struct crypto_async_request *creq, int err)
|
|
{
|
|
struct swcr_req *req = creq->data;
|
|
|
|
dprintk("%s()\n", __FUNCTION__);
|
|
if (err) {
|
|
if (err == -EINPROGRESS)
|
|
return;
|
|
dprintk("%s() fail %d\n", __FUNCTION__, -err);
|
|
req->crp->crp_etype = -err;
|
|
}
|
|
|
|
swcr_process_req_complete(req);
|
|
}
|
|
#endif /* defined(HAVE_ABLKCIPHER) || defined(HAVE_AHASH) */
|
|
|
|
|
|
static void swcr_process_req(struct swcr_req *req)
|
|
{
|
|
struct swcr_data *sw;
|
|
struct cryptop *crp = req->crp;
|
|
struct cryptodesc *crd = req->crd;
|
|
struct sk_buff *skb = (struct sk_buff *) crp->crp_buf;
|
|
struct uio *uiop = (struct uio *) crp->crp_buf;
|
|
int sg_num, sg_len, skip;
|
|
|
|
dprintk("%s()\n", __FUNCTION__);
|
|
|
|
/*
|
|
* Find the crypto context.
|
|
*
|
|
* XXX Note that the logic here prevents us from having
|
|
* XXX the same algorithm multiple times in a session
|
|
* XXX (or rather, we can but it won't give us the right
|
|
* XXX results). To do that, we'd need some way of differentiating
|
|
* XXX between the various instances of an algorithm (so we can
|
|
* XXX locate the correct crypto context).
|
|
*/
|
|
for (sw = req->sw_head; sw && sw->sw_alg != crd->crd_alg; sw = sw->sw_next)
|
|
;
|
|
|
|
/* No such context ? */
|
|
if (sw == NULL) {
|
|
crp->crp_etype = EINVAL;
|
|
dprintk("%s,%d: EINVAL\n", __FILE__, __LINE__);
|
|
goto done;
|
|
}
|
|
|
|
/*
|
|
* for some types we need to ensure only one user as info is stored in
|
|
* the tfm during an operation that can get corrupted
|
|
*/
|
|
switch (sw->sw_type & SW_TYPE_ALG_AMASK) {
|
|
#ifdef HAVE_AHASH
|
|
case SW_TYPE_AHMAC:
|
|
case SW_TYPE_AHASH:
|
|
#endif
|
|
case SW_TYPE_HMAC:
|
|
case SW_TYPE_HASH: {
|
|
unsigned long flags;
|
|
spin_lock_irqsave(&sw->sw_tfm_lock, flags);
|
|
if (sw->sw_type & SW_TYPE_INUSE) {
|
|
spin_unlock_irqrestore(&sw->sw_tfm_lock, flags);
|
|
execute_later((void (*)(void *))swcr_process_req, (void *)req);
|
|
return;
|
|
}
|
|
sw->sw_type |= SW_TYPE_INUSE;
|
|
spin_unlock_irqrestore(&sw->sw_tfm_lock, flags);
|
|
} break;
|
|
}
|
|
|
|
req->sw = sw;
|
|
skip = crd->crd_skip;
|
|
|
|
/*
|
|
* setup the SG list skip from the start of the buffer
|
|
*/
|
|
memset(req->sg, 0, sizeof(req->sg));
|
|
sg_init_table(req->sg, SCATTERLIST_MAX);
|
|
if (crp->crp_flags & CRYPTO_F_SKBUF) {
|
|
int i, len;
|
|
|
|
sg_num = 0;
|
|
sg_len = 0;
|
|
|
|
if (skip < skb_headlen(skb)) {
|
|
len = skb_headlen(skb) - skip;
|
|
if (len + sg_len > crd->crd_len)
|
|
len = crd->crd_len - sg_len;
|
|
sg_set_page(&req->sg[sg_num],
|
|
virt_to_page(skb->data + skip), len,
|
|
offset_in_page(skb->data + skip));
|
|
sg_len += len;
|
|
sg_num++;
|
|
skip = 0;
|
|
} else
|
|
skip -= skb_headlen(skb);
|
|
|
|
for (i = 0; sg_len < crd->crd_len &&
|
|
i < skb_shinfo(skb)->nr_frags &&
|
|
sg_num < SCATTERLIST_MAX; i++) {
|
|
if (skip < skb_shinfo(skb)->frags[i].size) {
|
|
len = skb_shinfo(skb)->frags[i].size - skip;
|
|
if (len + sg_len > crd->crd_len)
|
|
len = crd->crd_len - sg_len;
|
|
sg_set_page(&req->sg[sg_num],
|
|
skb_frag_page(&skb_shinfo(skb)->frags[i]),
|
|
len,
|
|
skb_shinfo(skb)->frags[i].page_offset + skip);
|
|
sg_len += len;
|
|
sg_num++;
|
|
skip = 0;
|
|
} else
|
|
skip -= skb_shinfo(skb)->frags[i].size;
|
|
}
|
|
} else if (crp->crp_flags & CRYPTO_F_IOV) {
|
|
int len;
|
|
|
|
sg_len = 0;
|
|
for (sg_num = 0; sg_len < crd->crd_len &&
|
|
sg_num < uiop->uio_iovcnt &&
|
|
sg_num < SCATTERLIST_MAX; sg_num++) {
|
|
if (skip <= uiop->uio_iov[sg_num].iov_len) {
|
|
len = uiop->uio_iov[sg_num].iov_len - skip;
|
|
if (len + sg_len > crd->crd_len)
|
|
len = crd->crd_len - sg_len;
|
|
sg_set_page(&req->sg[sg_num],
|
|
virt_to_page(uiop->uio_iov[sg_num].iov_base+skip),
|
|
len,
|
|
offset_in_page(uiop->uio_iov[sg_num].iov_base+skip));
|
|
sg_len += len;
|
|
skip = 0;
|
|
} else
|
|
skip -= uiop->uio_iov[sg_num].iov_len;
|
|
}
|
|
} else {
|
|
sg_len = (crp->crp_ilen - skip);
|
|
if (sg_len > crd->crd_len)
|
|
sg_len = crd->crd_len;
|
|
sg_set_page(&req->sg[0], virt_to_page(crp->crp_buf + skip),
|
|
sg_len, offset_in_page(crp->crp_buf + skip));
|
|
sg_num = 1;
|
|
}
|
|
if (sg_num > 0)
|
|
sg_mark_end(&req->sg[sg_num-1]);
|
|
|
|
switch (sw->sw_type & SW_TYPE_ALG_AMASK) {
|
|
|
|
#ifdef HAVE_AHASH
|
|
case SW_TYPE_AHMAC:
|
|
case SW_TYPE_AHASH:
|
|
{
|
|
int ret;
|
|
|
|
/* check we have room for the result */
|
|
if (crp->crp_ilen - crd->crd_inject < sw->u.hmac.sw_mlen) {
|
|
dprintk("cryptosoft: EINVAL crp_ilen=%d, len=%d, inject=%d "
|
|
"digestsize=%d\n", crp->crp_ilen, crd->crd_skip + sg_len,
|
|
crd->crd_inject, sw->u.hmac.sw_mlen);
|
|
crp->crp_etype = EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
req->crypto_req =
|
|
ahash_request_alloc(__crypto_ahash_cast(sw->sw_tfm),GFP_ATOMIC);
|
|
if (!req->crypto_req) {
|
|
crp->crp_etype = ENOMEM;
|
|
dprintk("%s,%d: ENOMEM ahash_request_alloc", __FILE__, __LINE__);
|
|
goto done;
|
|
}
|
|
|
|
ahash_request_set_callback(req->crypto_req,
|
|
CRYPTO_TFM_REQ_MAY_BACKLOG, swcr_process_callback, req);
|
|
|
|
memset(req->result, 0, sizeof(req->result));
|
|
|
|
if (sw->sw_type & SW_TYPE_AHMAC)
|
|
crypto_ahash_setkey(__crypto_ahash_cast(sw->sw_tfm),
|
|
sw->u.hmac.sw_key, sw->u.hmac.sw_klen);
|
|
ahash_request_set_crypt(req->crypto_req, req->sg, req->result, sg_len);
|
|
ret = crypto_ahash_digest(req->crypto_req);
|
|
switch (ret) {
|
|
case -EINPROGRESS:
|
|
case -EBUSY:
|
|
return;
|
|
default:
|
|
case 0:
|
|
dprintk("hash OP %s %d\n", ret ? "failed" : "success", ret);
|
|
crp->crp_etype = ret;
|
|
goto done;
|
|
}
|
|
} break;
|
|
#endif /* HAVE_AHASH */
|
|
|
|
#ifdef HAVE_ABLKCIPHER
|
|
case SW_TYPE_ABLKCIPHER: {
|
|
int ret;
|
|
unsigned char *ivp = req->iv;
|
|
int ivsize =
|
|
crypto_ablkcipher_ivsize(__crypto_ablkcipher_cast(sw->sw_tfm));
|
|
|
|
if (sg_len < crypto_ablkcipher_blocksize(
|
|
__crypto_ablkcipher_cast(sw->sw_tfm))) {
|
|
crp->crp_etype = EINVAL;
|
|
dprintk("%s,%d: EINVAL len %d < %d\n", __FILE__, __LINE__,
|
|
sg_len, crypto_ablkcipher_blocksize(
|
|
__crypto_ablkcipher_cast(sw->sw_tfm)));
|
|
goto done;
|
|
}
|
|
|
|
if (ivsize > sizeof(req->iv)) {
|
|
crp->crp_etype = EINVAL;
|
|
dprintk("%s,%d: EINVAL\n", __FILE__, __LINE__);
|
|
goto done;
|
|
}
|
|
|
|
req->crypto_req = ablkcipher_request_alloc(
|
|
__crypto_ablkcipher_cast(sw->sw_tfm), GFP_ATOMIC);
|
|
if (!req->crypto_req) {
|
|
crp->crp_etype = ENOMEM;
|
|
dprintk("%s,%d: ENOMEM ablkcipher_request_alloc",
|
|
__FILE__, __LINE__);
|
|
goto done;
|
|
}
|
|
|
|
ablkcipher_request_set_callback(req->crypto_req,
|
|
CRYPTO_TFM_REQ_MAY_BACKLOG, swcr_process_callback, req);
|
|
|
|
if (crd->crd_flags & CRD_F_KEY_EXPLICIT) {
|
|
int i, error;
|
|
|
|
if (debug) {
|
|
dprintk("%s key:", __FUNCTION__);
|
|
for (i = 0; i < (crd->crd_klen + 7) / 8; i++)
|
|
dprintk("%s0x%x", (i % 8) ? " " : "\n ",
|
|
crd->crd_key[i] & 0xff);
|
|
dprintk("\n");
|
|
}
|
|
/* OCF doesn't enforce keys */
|
|
crypto_ablkcipher_set_flags(__crypto_ablkcipher_cast(sw->sw_tfm),
|
|
CRYPTO_TFM_REQ_WEAK_KEY);
|
|
error = crypto_ablkcipher_setkey(
|
|
__crypto_ablkcipher_cast(sw->sw_tfm), crd->crd_key,
|
|
(crd->crd_klen + 7) / 8);
|
|
if (error) {
|
|
dprintk("cryptosoft: setkey failed %d (crt_flags=0x%x)\n",
|
|
error, sw->sw_tfm->crt_flags);
|
|
crp->crp_etype = -error;
|
|
}
|
|
}
|
|
|
|
if (crd->crd_flags & CRD_F_ENCRYPT) { /* encrypt */
|
|
|
|
if (crd->crd_flags & CRD_F_IV_EXPLICIT)
|
|
ivp = crd->crd_iv;
|
|
else
|
|
get_random_bytes(ivp, ivsize);
|
|
/*
|
|
* do we have to copy the IV back to the buffer ?
|
|
*/
|
|
if ((crd->crd_flags & CRD_F_IV_PRESENT) == 0) {
|
|
crypto_copyback(crp->crp_flags, crp->crp_buf,
|
|
crd->crd_inject, ivsize, (caddr_t)ivp);
|
|
}
|
|
ablkcipher_request_set_crypt(req->crypto_req, req->sg, req->sg,
|
|
sg_len, ivp);
|
|
ret = crypto_ablkcipher_encrypt(req->crypto_req);
|
|
|
|
} else { /*decrypt */
|
|
|
|
if (crd->crd_flags & CRD_F_IV_EXPLICIT)
|
|
ivp = crd->crd_iv;
|
|
else
|
|
crypto_copydata(crp->crp_flags, crp->crp_buf,
|
|
crd->crd_inject, ivsize, (caddr_t)ivp);
|
|
ablkcipher_request_set_crypt(req->crypto_req, req->sg, req->sg,
|
|
sg_len, ivp);
|
|
ret = crypto_ablkcipher_decrypt(req->crypto_req);
|
|
}
|
|
|
|
switch (ret) {
|
|
case -EINPROGRESS:
|
|
case -EBUSY:
|
|
return;
|
|
default:
|
|
case 0:
|
|
dprintk("crypto OP %s %d\n", ret ? "failed" : "success", ret);
|
|
crp->crp_etype = ret;
|
|
goto done;
|
|
}
|
|
} break;
|
|
#endif /* HAVE_ABLKCIPHER */
|
|
|
|
case SW_TYPE_BLKCIPHER: {
|
|
unsigned char iv[EALG_MAX_BLOCK_LEN];
|
|
unsigned char *ivp = iv;
|
|
struct blkcipher_desc desc;
|
|
int ivsize = crypto_blkcipher_ivsize(crypto_blkcipher_cast(sw->sw_tfm));
|
|
|
|
if (sg_len < crypto_blkcipher_blocksize(
|
|
crypto_blkcipher_cast(sw->sw_tfm))) {
|
|
crp->crp_etype = EINVAL;
|
|
dprintk("%s,%d: EINVAL len %d < %d\n", __FILE__, __LINE__,
|
|
sg_len, crypto_blkcipher_blocksize(
|
|
crypto_blkcipher_cast(sw->sw_tfm)));
|
|
goto done;
|
|
}
|
|
|
|
if (ivsize > sizeof(iv)) {
|
|
crp->crp_etype = EINVAL;
|
|
dprintk("%s,%d: EINVAL\n", __FILE__, __LINE__);
|
|
goto done;
|
|
}
|
|
|
|
if (crd->crd_flags & CRD_F_KEY_EXPLICIT) {
|
|
int i, error;
|
|
|
|
if (debug) {
|
|
dprintk("%s key:", __FUNCTION__);
|
|
for (i = 0; i < (crd->crd_klen + 7) / 8; i++)
|
|
dprintk("%s0x%x", (i % 8) ? " " : "\n ",
|
|
crd->crd_key[i] & 0xff);
|
|
dprintk("\n");
|
|
}
|
|
/* OCF doesn't enforce keys */
|
|
crypto_blkcipher_set_flags(crypto_blkcipher_cast(sw->sw_tfm),
|
|
CRYPTO_TFM_REQ_WEAK_KEY);
|
|
error = crypto_blkcipher_setkey(
|
|
crypto_blkcipher_cast(sw->sw_tfm), crd->crd_key,
|
|
(crd->crd_klen + 7) / 8);
|
|
if (error) {
|
|
dprintk("cryptosoft: setkey failed %d (crt_flags=0x%x)\n",
|
|
error, sw->sw_tfm->crt_flags);
|
|
crp->crp_etype = -error;
|
|
}
|
|
}
|
|
|
|
memset(&desc, 0, sizeof(desc));
|
|
desc.tfm = crypto_blkcipher_cast(sw->sw_tfm);
|
|
|
|
if (crd->crd_flags & CRD_F_ENCRYPT) { /* encrypt */
|
|
|
|
if (crd->crd_flags & CRD_F_IV_EXPLICIT) {
|
|
ivp = crd->crd_iv;
|
|
} else {
|
|
get_random_bytes(ivp, ivsize);
|
|
}
|
|
/*
|
|
* do we have to copy the IV back to the buffer ?
|
|
*/
|
|
if ((crd->crd_flags & CRD_F_IV_PRESENT) == 0) {
|
|
crypto_copyback(crp->crp_flags, crp->crp_buf,
|
|
crd->crd_inject, ivsize, (caddr_t)ivp);
|
|
}
|
|
desc.info = ivp;
|
|
crypto_blkcipher_encrypt_iv(&desc, req->sg, req->sg, sg_len);
|
|
|
|
} else { /*decrypt */
|
|
|
|
if (crd->crd_flags & CRD_F_IV_EXPLICIT) {
|
|
ivp = crd->crd_iv;
|
|
} else {
|
|
crypto_copydata(crp->crp_flags, crp->crp_buf,
|
|
crd->crd_inject, ivsize, (caddr_t)ivp);
|
|
}
|
|
desc.info = ivp;
|
|
crypto_blkcipher_decrypt_iv(&desc, req->sg, req->sg, sg_len);
|
|
}
|
|
} break;
|
|
|
|
case SW_TYPE_HMAC:
|
|
case SW_TYPE_HASH:
|
|
{
|
|
char result[HASH_MAX_LEN];
|
|
struct hash_desc desc;
|
|
|
|
/* check we have room for the result */
|
|
if (crp->crp_ilen - crd->crd_inject < sw->u.hmac.sw_mlen) {
|
|
dprintk("cryptosoft: EINVAL crp_ilen=%d, len=%d, inject=%d "
|
|
"digestsize=%d\n", crp->crp_ilen, crd->crd_skip + sg_len,
|
|
crd->crd_inject, sw->u.hmac.sw_mlen);
|
|
crp->crp_etype = EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
memset(&desc, 0, sizeof(desc));
|
|
desc.tfm = crypto_hash_cast(sw->sw_tfm);
|
|
|
|
memset(result, 0, sizeof(result));
|
|
|
|
if (sw->sw_type & SW_TYPE_HMAC) {
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,19)
|
|
crypto_hmac(sw->sw_tfm, sw->u.hmac.sw_key, &sw->u.hmac.sw_klen,
|
|
req->sg, sg_num, result);
|
|
#else
|
|
crypto_hash_setkey(desc.tfm, sw->u.hmac.sw_key,
|
|
sw->u.hmac.sw_klen);
|
|
crypto_hash_digest(&desc, req->sg, sg_len, result);
|
|
#endif /* LINUX_VERSION_CODE < KERNEL_VERSION(2,6,19) */
|
|
|
|
} else { /* SW_TYPE_HASH */
|
|
crypto_hash_digest(&desc, req->sg, sg_len, result);
|
|
}
|
|
|
|
crypto_copyback(crp->crp_flags, crp->crp_buf,
|
|
crd->crd_inject, sw->u.hmac.sw_mlen, result);
|
|
}
|
|
break;
|
|
|
|
case SW_TYPE_COMP: {
|
|
void *ibuf = NULL;
|
|
void *obuf = sw->u.sw_comp_buf;
|
|
int ilen = sg_len, olen = CRYPTO_MAX_DATA_LEN;
|
|
int ret = 0;
|
|
|
|
/*
|
|
* we need to use an additional copy if there is more than one
|
|
* input chunk since the kernel comp routines do not handle
|
|
* SG yet. Otherwise we just use the input buffer as is.
|
|
* Rather than allocate another buffer we just split the tmp
|
|
* buffer we already have.
|
|
* Perhaps we should just use zlib directly ?
|
|
*/
|
|
if (sg_num > 1) {
|
|
int blk;
|
|
|
|
ibuf = obuf;
|
|
for (blk = 0; blk < sg_num; blk++) {
|
|
memcpy(obuf, sg_virt(&req->sg[blk]),
|
|
req->sg[blk].length);
|
|
obuf += req->sg[blk].length;
|
|
}
|
|
olen -= sg_len;
|
|
} else
|
|
ibuf = sg_virt(&req->sg[0]);
|
|
|
|
if (crd->crd_flags & CRD_F_ENCRYPT) { /* compress */
|
|
ret = crypto_comp_compress(crypto_comp_cast(sw->sw_tfm),
|
|
ibuf, ilen, obuf, &olen);
|
|
if (!ret && olen > crd->crd_len) {
|
|
dprintk("cryptosoft: ERANGE compress %d into %d\n",
|
|
crd->crd_len, olen);
|
|
if (swcr_fail_if_compression_grows)
|
|
ret = ERANGE;
|
|
}
|
|
} else { /* decompress */
|
|
ret = crypto_comp_decompress(crypto_comp_cast(sw->sw_tfm),
|
|
ibuf, ilen, obuf, &olen);
|
|
if (!ret && (olen + crd->crd_inject) > crp->crp_olen) {
|
|
dprintk("cryptosoft: ETOOSMALL decompress %d into %d, "
|
|
"space for %d,at offset %d\n",
|
|
crd->crd_len, olen, crp->crp_olen, crd->crd_inject);
|
|
ret = ETOOSMALL;
|
|
}
|
|
}
|
|
if (ret)
|
|
dprintk("%s,%d: ret = %d\n", __FILE__, __LINE__, ret);
|
|
|
|
/*
|
|
* on success copy result back,
|
|
* linux crpyto API returns -errno, we need to fix that
|
|
*/
|
|
crp->crp_etype = ret < 0 ? -ret : ret;
|
|
if (ret == 0) {
|
|
/* copy back the result and return it's size */
|
|
crypto_copyback(crp->crp_flags, crp->crp_buf,
|
|
crd->crd_inject, olen, obuf);
|
|
crp->crp_olen = olen;
|
|
}
|
|
} break;
|
|
|
|
default:
|
|
/* Unknown/unsupported algorithm */
|
|
dprintk("%s,%d: EINVAL\n", __FILE__, __LINE__);
|
|
crp->crp_etype = EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
done:
|
|
swcr_process_req_complete(req);
|
|
}
|
|
|
|
|
|
/*
|
|
* Process a crypto request.
|
|
*/
|
|
static int
|
|
swcr_process(device_t dev, struct cryptop *crp, int hint)
|
|
{
|
|
struct swcr_req *req = NULL;
|
|
u_int32_t lid;
|
|
|
|
dprintk("%s()\n", __FUNCTION__);
|
|
/* Sanity check */
|
|
if (crp == NULL) {
|
|
dprintk("%s,%d: EINVAL\n", __FILE__, __LINE__);
|
|
return EINVAL;
|
|
}
|
|
|
|
crp->crp_etype = 0;
|
|
|
|
if (crp->crp_desc == NULL || crp->crp_buf == NULL) {
|
|
dprintk("%s,%d: EINVAL\n", __FILE__, __LINE__);
|
|
crp->crp_etype = EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
lid = crp->crp_sid & 0xffffffff;
|
|
if (lid >= swcr_sesnum || lid == 0 || swcr_sessions == NULL ||
|
|
swcr_sessions[lid] == NULL) {
|
|
crp->crp_etype = ENOENT;
|
|
dprintk("%s,%d: ENOENT\n", __FILE__, __LINE__);
|
|
goto done;
|
|
}
|
|
|
|
/*
|
|
* do some error checking outside of the loop for SKB and IOV processing
|
|
* this leaves us with valid skb or uiop pointers for later
|
|
*/
|
|
if (crp->crp_flags & CRYPTO_F_SKBUF) {
|
|
struct sk_buff *skb = (struct sk_buff *) crp->crp_buf;
|
|
if (skb_shinfo(skb)->nr_frags >= SCATTERLIST_MAX) {
|
|
printk("%s,%d: %d nr_frags > SCATTERLIST_MAX", __FILE__, __LINE__,
|
|
skb_shinfo(skb)->nr_frags);
|
|
goto done;
|
|
}
|
|
} else if (crp->crp_flags & CRYPTO_F_IOV) {
|
|
struct uio *uiop = (struct uio *) crp->crp_buf;
|
|
if (uiop->uio_iovcnt > SCATTERLIST_MAX) {
|
|
printk("%s,%d: %d uio_iovcnt > SCATTERLIST_MAX", __FILE__, __LINE__,
|
|
uiop->uio_iovcnt);
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* setup a new request ready for queuing
|
|
*/
|
|
req = kmem_cache_alloc(swcr_req_cache, SLAB_ATOMIC);
|
|
if (req == NULL) {
|
|
dprintk("%s,%d: ENOMEM\n", __FILE__, __LINE__);
|
|
crp->crp_etype = ENOMEM;
|
|
goto done;
|
|
}
|
|
memset(req, 0, sizeof(*req));
|
|
|
|
req->sw_head = swcr_sessions[lid];
|
|
req->crp = crp;
|
|
req->crd = crp->crp_desc;
|
|
|
|
swcr_process_req(req);
|
|
return 0;
|
|
|
|
done:
|
|
crypto_done(crp);
|
|
if (req)
|
|
kmem_cache_free(swcr_req_cache, req);
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int
|
|
cryptosoft_init(void)
|
|
{
|
|
int i, sw_type, mode;
|
|
char *algo;
|
|
|
|
dprintk("%s(%p)\n", __FUNCTION__, cryptosoft_init);
|
|
|
|
swcr_req_cache = kmem_cache_create("cryptosoft_req",
|
|
sizeof(struct swcr_req), 0, SLAB_HWCACHE_ALIGN, NULL
|
|
#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,23)
|
|
, NULL
|
|
#endif
|
|
);
|
|
if (!swcr_req_cache) {
|
|
printk("cryptosoft: failed to create request cache\n");
|
|
return -ENOENT;
|
|
}
|
|
|
|
softc_device_init(&swcr_softc, "cryptosoft", 0, swcr_methods);
|
|
|
|
swcr_id = crypto_get_driverid(softc_get_device(&swcr_softc),
|
|
CRYPTOCAP_F_SOFTWARE | CRYPTOCAP_F_SYNC);
|
|
if (swcr_id < 0) {
|
|
printk("cryptosoft: Software crypto device cannot initialize!");
|
|
return -ENODEV;
|
|
}
|
|
|
|
#define REGISTER(alg) \
|
|
crypto_register(swcr_id, alg, 0,0)
|
|
|
|
for (i = 0; i < sizeof(crypto_details)/sizeof(crypto_details[0]); i++) {
|
|
int found;
|
|
|
|
algo = crypto_details[i].alg_name;
|
|
if (!algo || !*algo) {
|
|
dprintk("%s:Algorithm %d not supported\n", __FUNCTION__, i);
|
|
continue;
|
|
}
|
|
|
|
mode = crypto_details[i].mode;
|
|
sw_type = crypto_details[i].sw_type;
|
|
|
|
found = 0;
|
|
switch (sw_type & SW_TYPE_ALG_MASK) {
|
|
case SW_TYPE_CIPHER:
|
|
found = crypto_has_cipher(algo, 0, CRYPTO_ALG_ASYNC);
|
|
break;
|
|
case SW_TYPE_HMAC:
|
|
found = crypto_has_hash(algo, 0, swcr_no_ahash?CRYPTO_ALG_ASYNC:0);
|
|
break;
|
|
case SW_TYPE_HASH:
|
|
found = crypto_has_hash(algo, 0, swcr_no_ahash?CRYPTO_ALG_ASYNC:0);
|
|
break;
|
|
case SW_TYPE_COMP:
|
|
found = crypto_has_comp(algo, 0, CRYPTO_ALG_ASYNC);
|
|
break;
|
|
case SW_TYPE_BLKCIPHER:
|
|
found = crypto_has_blkcipher(algo, 0, CRYPTO_ALG_ASYNC);
|
|
if (!found && !swcr_no_ablk)
|
|
found = crypto_has_ablkcipher(algo, 0, 0);
|
|
break;
|
|
}
|
|
if (found) {
|
|
REGISTER(i);
|
|
} else {
|
|
dprintk("%s:Algorithm Type %d not supported (algorithm %d:'%s')\n",
|
|
__FUNCTION__, sw_type, i, algo);
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
cryptosoft_exit(void)
|
|
{
|
|
dprintk("%s()\n", __FUNCTION__);
|
|
crypto_unregister_all(swcr_id);
|
|
swcr_id = -1;
|
|
kmem_cache_destroy(swcr_req_cache);
|
|
}
|
|
|
|
late_initcall(cryptosoft_init);
|
|
module_exit(cryptosoft_exit);
|
|
|
|
MODULE_LICENSE("Dual BSD/GPL");
|
|
MODULE_AUTHOR("David McCullough <david_mccullough@mcafee.com>");
|
|
MODULE_DESCRIPTION("Cryptosoft (OCF module for kernel crypto)");
|