summaryrefslogtreecommitdiff
path: root/crypto
diff options
context:
space:
mode:
authorSzilveszter Ördög <slipszi@gmail.com>2010-08-06 05:26:38 +0400
committerHerbert Xu <herbert@gondor.apana.org.au>2010-08-06 05:26:38 +0400
commit23a75eee070f1370bee803a34f285cf81eb5f331 (patch)
tree6427c53a261840661f135b99d81062fc015dd571 /crypto
parentfc1caf6eafb30ea185720e29f7f5eccca61ecd60 (diff)
downloadlinux-23a75eee070f1370bee803a34f285cf81eb5f331.tar.xz
crypto: hash - Fix handling of small unaligned buffers
If a scatterwalk chain contains an entry with an unaligned offset then hash_walk_next() will cut off the next step at the next alignment point. However, if the entry ends before the next alignment point then we a loop, which leads to a kernel oops. Fix this by checking whether the next aligment point is before the end of the current entry. Signed-off-by: Szilveszter Ördög <slipszi@gmail.com> Acked-by: David S. Miller <davem@davemloft.net> Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
Diffstat (limited to 'crypto')
-rw-r--r--crypto/ahash.c7
1 files changed, 5 insertions, 2 deletions
diff --git a/crypto/ahash.c b/crypto/ahash.c
index b8c59b889c6e..f669822a7a44 100644
--- a/crypto/ahash.c
+++ b/crypto/ahash.c
@@ -47,8 +47,11 @@ static int hash_walk_next(struct crypto_hash_walk *walk)
walk->data = crypto_kmap(walk->pg, 0);
walk->data += offset;
- if (offset & alignmask)
- nbytes = alignmask + 1 - (offset & alignmask);
+ if (offset & alignmask) {
+ unsigned int unaligned = alignmask + 1 - (offset & alignmask);
+ if (nbytes > unaligned)
+ nbytes = unaligned;
+ }
walk->entrylen -= nbytes;
return nbytes;