diff options
Diffstat (limited to 'drivers/crypto/sahara.c')
-rw-r--r-- | drivers/crypto/sahara.c | 1070 |
1 files changed, 1070 insertions, 0 deletions
diff --git a/drivers/crypto/sahara.c b/drivers/crypto/sahara.c new file mode 100644 index 000000000000..a97bb6c1596c --- /dev/null +++ b/drivers/crypto/sahara.c | |||
@@ -0,0 +1,1070 @@ | |||
1 | /* | ||
2 | * Cryptographic API. | ||
3 | * | ||
4 | * Support for SAHARA cryptographic accelerator. | ||
5 | * | ||
6 | * Copyright (c) 2013 Vista Silicon S.L. | ||
7 | * Author: Javier Martin <javier.martin@vista-silicon.com> | ||
8 | * | ||
9 | * This program is free software; you can redistribute it and/or modify | ||
10 | * it under the terms of the GNU General Public License version 2 as published | ||
11 | * by the Free Software Foundation. | ||
12 | * | ||
13 | * Based on omap-aes.c and tegra-aes.c | ||
14 | */ | ||
15 | |||
16 | #include <crypto/algapi.h> | ||
17 | #include <crypto/aes.h> | ||
18 | |||
19 | #include <linux/clk.h> | ||
20 | #include <linux/crypto.h> | ||
21 | #include <linux/interrupt.h> | ||
22 | #include <linux/io.h> | ||
23 | #include <linux/irq.h> | ||
24 | #include <linux/kernel.h> | ||
25 | #include <linux/module.h> | ||
26 | #include <linux/of.h> | ||
27 | #include <linux/platform_device.h> | ||
28 | |||
29 | #define SAHARA_NAME "sahara" | ||
30 | #define SAHARA_VERSION_3 3 | ||
31 | #define SAHARA_TIMEOUT_MS 1000 | ||
32 | #define SAHARA_MAX_HW_DESC 2 | ||
33 | #define SAHARA_MAX_HW_LINK 20 | ||
34 | |||
35 | #define FLAGS_MODE_MASK 0x000f | ||
36 | #define FLAGS_ENCRYPT BIT(0) | ||
37 | #define FLAGS_CBC BIT(1) | ||
38 | #define FLAGS_NEW_KEY BIT(3) | ||
39 | #define FLAGS_BUSY 4 | ||
40 | |||
41 | #define SAHARA_HDR_BASE 0x00800000 | ||
42 | #define SAHARA_HDR_SKHA_ALG_AES 0 | ||
43 | #define SAHARA_HDR_SKHA_OP_ENC (1 << 2) | ||
44 | #define SAHARA_HDR_SKHA_MODE_ECB (0 << 3) | ||
45 | #define SAHARA_HDR_SKHA_MODE_CBC (1 << 3) | ||
46 | #define SAHARA_HDR_FORM_DATA (5 << 16) | ||
47 | #define SAHARA_HDR_FORM_KEY (8 << 16) | ||
48 | #define SAHARA_HDR_LLO (1 << 24) | ||
49 | #define SAHARA_HDR_CHA_SKHA (1 << 28) | ||
50 | #define SAHARA_HDR_CHA_MDHA (2 << 28) | ||
51 | #define SAHARA_HDR_PARITY_BIT (1 << 31) | ||
52 | |||
53 | /* SAHARA can only process one request at a time */ | ||
54 | #define SAHARA_QUEUE_LENGTH 1 | ||
55 | |||
56 | #define SAHARA_REG_VERSION 0x00 | ||
57 | #define SAHARA_REG_DAR 0x04 | ||
58 | #define SAHARA_REG_CONTROL 0x08 | ||
59 | #define SAHARA_CONTROL_SET_THROTTLE(x) (((x) & 0xff) << 24) | ||
60 | #define SAHARA_CONTROL_SET_MAXBURST(x) (((x) & 0xff) << 16) | ||
61 | #define SAHARA_CONTROL_RNG_AUTORSD (1 << 7) | ||
62 | #define SAHARA_CONTROL_ENABLE_INT (1 << 4) | ||
63 | #define SAHARA_REG_CMD 0x0C | ||
64 | #define SAHARA_CMD_RESET (1 << 0) | ||
65 | #define SAHARA_CMD_CLEAR_INT (1 << 8) | ||
66 | #define SAHARA_CMD_CLEAR_ERR (1 << 9) | ||
67 | #define SAHARA_CMD_SINGLE_STEP (1 << 10) | ||
68 | #define SAHARA_CMD_MODE_BATCH (1 << 16) | ||
69 | #define SAHARA_CMD_MODE_DEBUG (1 << 18) | ||
70 | #define SAHARA_REG_STATUS 0x10 | ||
71 | #define SAHARA_STATUS_GET_STATE(x) ((x) & 0x7) | ||
72 | #define SAHARA_STATE_IDLE 0 | ||
73 | #define SAHARA_STATE_BUSY 1 | ||
74 | #define SAHARA_STATE_ERR 2 | ||
75 | #define SAHARA_STATE_FAULT 3 | ||
76 | #define SAHARA_STATE_COMPLETE 4 | ||
77 | #define SAHARA_STATE_COMP_FLAG (1 << 2) | ||
78 | #define SAHARA_STATUS_DAR_FULL (1 << 3) | ||
79 | #define SAHARA_STATUS_ERROR (1 << 4) | ||
80 | #define SAHARA_STATUS_SECURE (1 << 5) | ||
81 | #define SAHARA_STATUS_FAIL (1 << 6) | ||
82 | #define SAHARA_STATUS_INIT (1 << 7) | ||
83 | #define SAHARA_STATUS_RNG_RESEED (1 << 8) | ||
84 | #define SAHARA_STATUS_ACTIVE_RNG (1 << 9) | ||
85 | #define SAHARA_STATUS_ACTIVE_MDHA (1 << 10) | ||
86 | #define SAHARA_STATUS_ACTIVE_SKHA (1 << 11) | ||
87 | #define SAHARA_STATUS_MODE_BATCH (1 << 16) | ||
88 | #define SAHARA_STATUS_MODE_DEDICATED (1 << 17) | ||
89 | #define SAHARA_STATUS_MODE_DEBUG (1 << 18) | ||
90 | #define SAHARA_STATUS_GET_ISTATE(x) (((x) >> 24) & 0xff) | ||
91 | #define SAHARA_REG_ERRSTATUS 0x14 | ||
92 | #define SAHARA_ERRSTATUS_GET_SOURCE(x) ((x) & 0xf) | ||
93 | #define SAHARA_ERRSOURCE_CHA 14 | ||
94 | #define SAHARA_ERRSOURCE_DMA 15 | ||
95 | #define SAHARA_ERRSTATUS_DMA_DIR (1 << 8) | ||
96 | #define SAHARA_ERRSTATUS_GET_DMASZ(x)(((x) >> 9) & 0x3) | ||
97 | #define SAHARA_ERRSTATUS_GET_DMASRC(x) (((x) >> 13) & 0x7) | ||
98 | #define SAHARA_ERRSTATUS_GET_CHASRC(x) (((x) >> 16) & 0xfff) | ||
99 | #define SAHARA_ERRSTATUS_GET_CHAERR(x) (((x) >> 28) & 0x3) | ||
100 | #define SAHARA_REG_FADDR 0x18 | ||
101 | #define SAHARA_REG_CDAR 0x1C | ||
102 | #define SAHARA_REG_IDAR 0x20 | ||
103 | |||
104 | struct sahara_hw_desc { | ||
105 | u32 hdr; | ||
106 | u32 len1; | ||
107 | dma_addr_t p1; | ||
108 | u32 len2; | ||
109 | dma_addr_t p2; | ||
110 | dma_addr_t next; | ||
111 | }; | ||
112 | |||
113 | struct sahara_hw_link { | ||
114 | u32 len; | ||
115 | dma_addr_t p; | ||
116 | dma_addr_t next; | ||
117 | }; | ||
118 | |||
119 | struct sahara_ctx { | ||
120 | struct sahara_dev *dev; | ||
121 | unsigned long flags; | ||
122 | int keylen; | ||
123 | u8 key[AES_KEYSIZE_128]; | ||
124 | struct crypto_ablkcipher *fallback; | ||
125 | }; | ||
126 | |||
127 | struct sahara_aes_reqctx { | ||
128 | unsigned long mode; | ||
129 | }; | ||
130 | |||
131 | struct sahara_dev { | ||
132 | struct device *device; | ||
133 | void __iomem *regs_base; | ||
134 | struct clk *clk_ipg; | ||
135 | struct clk *clk_ahb; | ||
136 | |||
137 | struct sahara_ctx *ctx; | ||
138 | spinlock_t lock; | ||
139 | struct crypto_queue queue; | ||
140 | unsigned long flags; | ||
141 | |||
142 | struct tasklet_struct done_task; | ||
143 | struct tasklet_struct queue_task; | ||
144 | |||
145 | struct sahara_hw_desc *hw_desc[SAHARA_MAX_HW_DESC]; | ||
146 | dma_addr_t hw_phys_desc[SAHARA_MAX_HW_DESC]; | ||
147 | |||
148 | u8 *key_base; | ||
149 | dma_addr_t key_phys_base; | ||
150 | |||
151 | u8 *iv_base; | ||
152 | dma_addr_t iv_phys_base; | ||
153 | |||
154 | struct sahara_hw_link *hw_link[SAHARA_MAX_HW_LINK]; | ||
155 | dma_addr_t hw_phys_link[SAHARA_MAX_HW_LINK]; | ||
156 | |||
157 | struct ablkcipher_request *req; | ||
158 | size_t total; | ||
159 | struct scatterlist *in_sg; | ||
160 | unsigned int nb_in_sg; | ||
161 | struct scatterlist *out_sg; | ||
162 | unsigned int nb_out_sg; | ||
163 | |||
164 | u32 error; | ||
165 | struct timer_list watchdog; | ||
166 | }; | ||
167 | |||
168 | static struct sahara_dev *dev_ptr; | ||
169 | |||
170 | static inline void sahara_write(struct sahara_dev *dev, u32 data, u32 reg) | ||
171 | { | ||
172 | writel(data, dev->regs_base + reg); | ||
173 | } | ||
174 | |||
175 | static inline unsigned int sahara_read(struct sahara_dev *dev, u32 reg) | ||
176 | { | ||
177 | return readl(dev->regs_base + reg); | ||
178 | } | ||
179 | |||
180 | static u32 sahara_aes_key_hdr(struct sahara_dev *dev) | ||
181 | { | ||
182 | u32 hdr = SAHARA_HDR_BASE | SAHARA_HDR_SKHA_ALG_AES | | ||
183 | SAHARA_HDR_FORM_KEY | SAHARA_HDR_LLO | | ||
184 | SAHARA_HDR_CHA_SKHA | SAHARA_HDR_PARITY_BIT; | ||
185 | |||
186 | if (dev->flags & FLAGS_CBC) { | ||
187 | hdr |= SAHARA_HDR_SKHA_MODE_CBC; | ||
188 | hdr ^= SAHARA_HDR_PARITY_BIT; | ||
189 | } | ||
190 | |||
191 | if (dev->flags & FLAGS_ENCRYPT) { | ||
192 | hdr |= SAHARA_HDR_SKHA_OP_ENC; | ||
193 | hdr ^= SAHARA_HDR_PARITY_BIT; | ||
194 | } | ||
195 | |||
196 | return hdr; | ||
197 | } | ||
198 | |||
199 | static u32 sahara_aes_data_link_hdr(struct sahara_dev *dev) | ||
200 | { | ||
201 | return SAHARA_HDR_BASE | SAHARA_HDR_FORM_DATA | | ||
202 | SAHARA_HDR_CHA_SKHA | SAHARA_HDR_PARITY_BIT; | ||
203 | } | ||
204 | |||
205 | static int sahara_sg_length(struct scatterlist *sg, | ||
206 | unsigned int total) | ||
207 | { | ||
208 | int sg_nb; | ||
209 | unsigned int len; | ||
210 | struct scatterlist *sg_list; | ||
211 | |||
212 | sg_nb = 0; | ||
213 | sg_list = sg; | ||
214 | |||
215 | while (total) { | ||
216 | len = min(sg_list->length, total); | ||
217 | |||
218 | sg_nb++; | ||
219 | total -= len; | ||
220 | |||
221 | sg_list = sg_next(sg_list); | ||
222 | if (!sg_list) | ||
223 | total = 0; | ||
224 | } | ||
225 | |||
226 | return sg_nb; | ||
227 | } | ||
228 | |||
229 | static char *sahara_err_src[16] = { | ||
230 | "No error", | ||
231 | "Header error", | ||
232 | "Descriptor length error", | ||
233 | "Descriptor length or pointer error", | ||
234 | "Link length error", | ||
235 | "Link pointer error", | ||
236 | "Input buffer error", | ||
237 | "Output buffer error", | ||
238 | "Output buffer starvation", | ||
239 | "Internal state fault", | ||
240 | "General descriptor problem", | ||
241 | "Reserved", | ||
242 | "Descriptor address error", | ||
243 | "Link address error", | ||
244 | "CHA error", | ||
245 | "DMA error" | ||
246 | }; | ||
247 | |||
248 | static char *sahara_err_dmasize[4] = { | ||
249 | "Byte transfer", | ||
250 | "Half-word transfer", | ||
251 | "Word transfer", | ||
252 | "Reserved" | ||
253 | }; | ||
254 | |||
255 | static char *sahara_err_dmasrc[8] = { | ||
256 | "No error", | ||
257 | "AHB bus error", | ||
258 | "Internal IP bus error", | ||
259 | "Parity error", | ||
260 | "DMA crosses 256 byte boundary", | ||
261 | "DMA is busy", | ||
262 | "Reserved", | ||
263 | "DMA HW error" | ||
264 | }; | ||
265 | |||
266 | static char *sahara_cha_errsrc[12] = { | ||
267 | "Input buffer non-empty", | ||
268 | "Illegal address", | ||
269 | "Illegal mode", | ||
270 | "Illegal data size", | ||
271 | "Illegal key size", | ||
272 | "Write during processing", | ||
273 | "CTX read during processing", | ||
274 | "HW error", | ||
275 | "Input buffer disabled/underflow", | ||
276 | "Output buffer disabled/overflow", | ||
277 | "DES key parity error", | ||
278 | "Reserved" | ||
279 | }; | ||
280 | |||
281 | static char *sahara_cha_err[4] = { "No error", "SKHA", "MDHA", "RNG" }; | ||
282 | |||
283 | static void sahara_decode_error(struct sahara_dev *dev, unsigned int error) | ||
284 | { | ||
285 | u8 source = SAHARA_ERRSTATUS_GET_SOURCE(error); | ||
286 | u16 chasrc = ffs(SAHARA_ERRSTATUS_GET_CHASRC(error)); | ||
287 | |||
288 | dev_err(dev->device, "%s: Error Register = 0x%08x\n", __func__, error); | ||
289 | |||
290 | dev_err(dev->device, " - %s.\n", sahara_err_src[source]); | ||
291 | |||
292 | if (source == SAHARA_ERRSOURCE_DMA) { | ||
293 | if (error & SAHARA_ERRSTATUS_DMA_DIR) | ||
294 | dev_err(dev->device, " * DMA read.\n"); | ||
295 | else | ||
296 | dev_err(dev->device, " * DMA write.\n"); | ||
297 | |||
298 | dev_err(dev->device, " * %s.\n", | ||
299 | sahara_err_dmasize[SAHARA_ERRSTATUS_GET_DMASZ(error)]); | ||
300 | dev_err(dev->device, " * %s.\n", | ||
301 | sahara_err_dmasrc[SAHARA_ERRSTATUS_GET_DMASRC(error)]); | ||
302 | } else if (source == SAHARA_ERRSOURCE_CHA) { | ||
303 | dev_err(dev->device, " * %s.\n", | ||
304 | sahara_cha_errsrc[chasrc]); | ||
305 | dev_err(dev->device, " * %s.\n", | ||
306 | sahara_cha_err[SAHARA_ERRSTATUS_GET_CHAERR(error)]); | ||
307 | } | ||
308 | dev_err(dev->device, "\n"); | ||
309 | } | ||
310 | |||
311 | static char *sahara_state[4] = { "Idle", "Busy", "Error", "HW Fault" }; | ||
312 | |||
313 | static void sahara_decode_status(struct sahara_dev *dev, unsigned int status) | ||
314 | { | ||
315 | u8 state; | ||
316 | |||
317 | if (!IS_ENABLED(DEBUG)) | ||
318 | return; | ||
319 | |||
320 | state = SAHARA_STATUS_GET_STATE(status); | ||
321 | |||
322 | dev_dbg(dev->device, "%s: Status Register = 0x%08x\n", | ||
323 | __func__, status); | ||
324 | |||
325 | dev_dbg(dev->device, " - State = %d:\n", state); | ||
326 | if (state & SAHARA_STATE_COMP_FLAG) | ||
327 | dev_dbg(dev->device, " * Descriptor completed. IRQ pending.\n"); | ||
328 | |||
329 | dev_dbg(dev->device, " * %s.\n", | ||
330 | sahara_state[state & ~SAHARA_STATE_COMP_FLAG]); | ||
331 | |||
332 | if (status & SAHARA_STATUS_DAR_FULL) | ||
333 | dev_dbg(dev->device, " - DAR Full.\n"); | ||
334 | if (status & SAHARA_STATUS_ERROR) | ||
335 | dev_dbg(dev->device, " - Error.\n"); | ||
336 | if (status & SAHARA_STATUS_SECURE) | ||
337 | dev_dbg(dev->device, " - Secure.\n"); | ||
338 | if (status & SAHARA_STATUS_FAIL) | ||
339 | dev_dbg(dev->device, " - Fail.\n"); | ||
340 | if (status & SAHARA_STATUS_RNG_RESEED) | ||
341 | dev_dbg(dev->device, " - RNG Reseed Request.\n"); | ||
342 | if (status & SAHARA_STATUS_ACTIVE_RNG) | ||
343 | dev_dbg(dev->device, " - RNG Active.\n"); | ||
344 | if (status & SAHARA_STATUS_ACTIVE_MDHA) | ||
345 | dev_dbg(dev->device, " - MDHA Active.\n"); | ||
346 | if (status & SAHARA_STATUS_ACTIVE_SKHA) | ||
347 | dev_dbg(dev->device, " - SKHA Active.\n"); | ||
348 | |||
349 | if (status & SAHARA_STATUS_MODE_BATCH) | ||
350 | dev_dbg(dev->device, " - Batch Mode.\n"); | ||
351 | else if (status & SAHARA_STATUS_MODE_DEDICATED) | ||
352 | dev_dbg(dev->device, " - Decidated Mode.\n"); | ||
353 | else if (status & SAHARA_STATUS_MODE_DEBUG) | ||
354 | dev_dbg(dev->device, " - Debug Mode.\n"); | ||
355 | |||
356 | dev_dbg(dev->device, " - Internal state = 0x%02x\n", | ||
357 | SAHARA_STATUS_GET_ISTATE(status)); | ||
358 | |||
359 | dev_dbg(dev->device, "Current DAR: 0x%08x\n", | ||
360 | sahara_read(dev, SAHARA_REG_CDAR)); | ||
361 | dev_dbg(dev->device, "Initial DAR: 0x%08x\n\n", | ||
362 | sahara_read(dev, SAHARA_REG_IDAR)); | ||
363 | } | ||
364 | |||
365 | static void sahara_dump_descriptors(struct sahara_dev *dev) | ||
366 | { | ||
367 | int i; | ||
368 | |||
369 | if (!IS_ENABLED(DEBUG)) | ||
370 | return; | ||
371 | |||
372 | for (i = 0; i < SAHARA_MAX_HW_DESC; i++) { | ||
373 | dev_dbg(dev->device, "Descriptor (%d) (0x%08x):\n", | ||
374 | i, dev->hw_phys_desc[i]); | ||
375 | dev_dbg(dev->device, "\thdr = 0x%08x\n", dev->hw_desc[i]->hdr); | ||
376 | dev_dbg(dev->device, "\tlen1 = %u\n", dev->hw_desc[i]->len1); | ||
377 | dev_dbg(dev->device, "\tp1 = 0x%08x\n", dev->hw_desc[i]->p1); | ||
378 | dev_dbg(dev->device, "\tlen2 = %u\n", dev->hw_desc[i]->len2); | ||
379 | dev_dbg(dev->device, "\tp2 = 0x%08x\n", dev->hw_desc[i]->p2); | ||
380 | dev_dbg(dev->device, "\tnext = 0x%08x\n", | ||
381 | dev->hw_desc[i]->next); | ||
382 | } | ||
383 | dev_dbg(dev->device, "\n"); | ||
384 | } | ||
385 | |||
386 | static void sahara_dump_links(struct sahara_dev *dev) | ||
387 | { | ||
388 | int i; | ||
389 | |||
390 | if (!IS_ENABLED(DEBUG)) | ||
391 | return; | ||
392 | |||
393 | for (i = 0; i < SAHARA_MAX_HW_LINK; i++) { | ||
394 | dev_dbg(dev->device, "Link (%d) (0x%08x):\n", | ||
395 | i, dev->hw_phys_link[i]); | ||
396 | dev_dbg(dev->device, "\tlen = %u\n", dev->hw_link[i]->len); | ||
397 | dev_dbg(dev->device, "\tp = 0x%08x\n", dev->hw_link[i]->p); | ||
398 | dev_dbg(dev->device, "\tnext = 0x%08x\n", | ||
399 | dev->hw_link[i]->next); | ||
400 | } | ||
401 | dev_dbg(dev->device, "\n"); | ||
402 | } | ||
403 | |||
404 | static void sahara_aes_done_task(unsigned long data) | ||
405 | { | ||
406 | struct sahara_dev *dev = (struct sahara_dev *)data; | ||
407 | |||
408 | dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, | ||
409 | DMA_TO_DEVICE); | ||
410 | dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, | ||
411 | DMA_FROM_DEVICE); | ||
412 | |||
413 | spin_lock(&dev->lock); | ||
414 | clear_bit(FLAGS_BUSY, &dev->flags); | ||
415 | spin_unlock(&dev->lock); | ||
416 | |||
417 | dev->req->base.complete(&dev->req->base, dev->error); | ||
418 | } | ||
419 | |||
420 | void sahara_watchdog(unsigned long data) | ||
421 | { | ||
422 | struct sahara_dev *dev = (struct sahara_dev *)data; | ||
423 | unsigned int err = sahara_read(dev, SAHARA_REG_ERRSTATUS); | ||
424 | unsigned int stat = sahara_read(dev, SAHARA_REG_STATUS); | ||
425 | |||
426 | sahara_decode_status(dev, stat); | ||
427 | sahara_decode_error(dev, err); | ||
428 | dev->error = -ETIMEDOUT; | ||
429 | sahara_aes_done_task(data); | ||
430 | } | ||
431 | |||
432 | static int sahara_hw_descriptor_create(struct sahara_dev *dev) | ||
433 | { | ||
434 | struct sahara_ctx *ctx = dev->ctx; | ||
435 | struct scatterlist *sg; | ||
436 | int ret; | ||
437 | int i, j; | ||
438 | |||
439 | /* Copy new key if necessary */ | ||
440 | if (ctx->flags & FLAGS_NEW_KEY) { | ||
441 | memcpy(dev->key_base, ctx->key, ctx->keylen); | ||
442 | ctx->flags &= ~FLAGS_NEW_KEY; | ||
443 | |||
444 | if (dev->flags & FLAGS_CBC) { | ||
445 | dev->hw_desc[0]->len1 = AES_BLOCK_SIZE; | ||
446 | dev->hw_desc[0]->p1 = dev->iv_phys_base; | ||
447 | } else { | ||
448 | dev->hw_desc[0]->len1 = 0; | ||
449 | dev->hw_desc[0]->p1 = 0; | ||
450 | } | ||
451 | dev->hw_desc[0]->len2 = ctx->keylen; | ||
452 | dev->hw_desc[0]->p2 = dev->key_phys_base; | ||
453 | dev->hw_desc[0]->next = dev->hw_phys_desc[1]; | ||
454 | } | ||
455 | dev->hw_desc[0]->hdr = sahara_aes_key_hdr(dev); | ||
456 | |||
457 | dev->nb_in_sg = sahara_sg_length(dev->in_sg, dev->total); | ||
458 | dev->nb_out_sg = sahara_sg_length(dev->out_sg, dev->total); | ||
459 | if ((dev->nb_in_sg + dev->nb_out_sg) > SAHARA_MAX_HW_LINK) { | ||
460 | dev_err(dev->device, "not enough hw links (%d)\n", | ||
461 | dev->nb_in_sg + dev->nb_out_sg); | ||
462 | return -EINVAL; | ||
463 | } | ||
464 | |||
465 | ret = dma_map_sg(dev->device, dev->in_sg, dev->nb_in_sg, | ||
466 | DMA_TO_DEVICE); | ||
467 | if (ret != dev->nb_in_sg) { | ||
468 | dev_err(dev->device, "couldn't map in sg\n"); | ||
469 | goto unmap_in; | ||
470 | } | ||
471 | ret = dma_map_sg(dev->device, dev->out_sg, dev->nb_out_sg, | ||
472 | DMA_FROM_DEVICE); | ||
473 | if (ret != dev->nb_out_sg) { | ||
474 | dev_err(dev->device, "couldn't map out sg\n"); | ||
475 | goto unmap_out; | ||
476 | } | ||
477 | |||
478 | /* Create input links */ | ||
479 | dev->hw_desc[1]->p1 = dev->hw_phys_link[0]; | ||
480 | sg = dev->in_sg; | ||
481 | for (i = 0; i < dev->nb_in_sg; i++) { | ||
482 | dev->hw_link[i]->len = sg->length; | ||
483 | dev->hw_link[i]->p = sg->dma_address; | ||
484 | if (i == (dev->nb_in_sg - 1)) { | ||
485 | dev->hw_link[i]->next = 0; | ||
486 | } else { | ||
487 | dev->hw_link[i]->next = dev->hw_phys_link[i + 1]; | ||
488 | sg = sg_next(sg); | ||
489 | } | ||
490 | } | ||
491 | |||
492 | /* Create output links */ | ||
493 | dev->hw_desc[1]->p2 = dev->hw_phys_link[i]; | ||
494 | sg = dev->out_sg; | ||
495 | for (j = i; j < dev->nb_out_sg + i; j++) { | ||
496 | dev->hw_link[j]->len = sg->length; | ||
497 | dev->hw_link[j]->p = sg->dma_address; | ||
498 | if (j == (dev->nb_out_sg + i - 1)) { | ||
499 | dev->hw_link[j]->next = 0; | ||
500 | } else { | ||
501 | dev->hw_link[j]->next = dev->hw_phys_link[j + 1]; | ||
502 | sg = sg_next(sg); | ||
503 | } | ||
504 | } | ||
505 | |||
506 | /* Fill remaining fields of hw_desc[1] */ | ||
507 | dev->hw_desc[1]->hdr = sahara_aes_data_link_hdr(dev); | ||
508 | dev->hw_desc[1]->len1 = dev->total; | ||
509 | dev->hw_desc[1]->len2 = dev->total; | ||
510 | dev->hw_desc[1]->next = 0; | ||
511 | |||
512 | sahara_dump_descriptors(dev); | ||
513 | sahara_dump_links(dev); | ||
514 | |||
515 | /* Start processing descriptor chain. */ | ||
516 | mod_timer(&dev->watchdog, | ||
517 | jiffies + msecs_to_jiffies(SAHARA_TIMEOUT_MS)); | ||
518 | sahara_write(dev, dev->hw_phys_desc[0], SAHARA_REG_DAR); | ||
519 | |||
520 | return 0; | ||
521 | |||
522 | unmap_out: | ||
523 | dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, | ||
524 | DMA_TO_DEVICE); | ||
525 | unmap_in: | ||
526 | dma_unmap_sg(dev->device, dev->in_sg, dev->nb_in_sg, | ||
527 | DMA_FROM_DEVICE); | ||
528 | |||
529 | return -EINVAL; | ||
530 | } | ||
531 | |||
532 | static void sahara_aes_queue_task(unsigned long data) | ||
533 | { | ||
534 | struct sahara_dev *dev = (struct sahara_dev *)data; | ||
535 | struct crypto_async_request *async_req, *backlog; | ||
536 | struct sahara_ctx *ctx; | ||
537 | struct sahara_aes_reqctx *rctx; | ||
538 | struct ablkcipher_request *req; | ||
539 | int ret; | ||
540 | |||
541 | spin_lock(&dev->lock); | ||
542 | backlog = crypto_get_backlog(&dev->queue); | ||
543 | async_req = crypto_dequeue_request(&dev->queue); | ||
544 | if (!async_req) | ||
545 | clear_bit(FLAGS_BUSY, &dev->flags); | ||
546 | spin_unlock(&dev->lock); | ||
547 | |||
548 | if (!async_req) | ||
549 | return; | ||
550 | |||
551 | if (backlog) | ||
552 | backlog->complete(backlog, -EINPROGRESS); | ||
553 | |||
554 | req = ablkcipher_request_cast(async_req); | ||
555 | |||
556 | /* Request is ready to be dispatched by the device */ | ||
557 | dev_dbg(dev->device, | ||
558 | "dispatch request (nbytes=%d, src=%p, dst=%p)\n", | ||
559 | req->nbytes, req->src, req->dst); | ||
560 | |||
561 | /* assign new request to device */ | ||
562 | dev->req = req; | ||
563 | dev->total = req->nbytes; | ||
564 | dev->in_sg = req->src; | ||
565 | dev->out_sg = req->dst; | ||
566 | |||
567 | rctx = ablkcipher_request_ctx(req); | ||
568 | ctx = crypto_ablkcipher_ctx(crypto_ablkcipher_reqtfm(req)); | ||
569 | rctx->mode &= FLAGS_MODE_MASK; | ||
570 | dev->flags = (dev->flags & ~FLAGS_MODE_MASK) | rctx->mode; | ||
571 | |||
572 | if ((dev->flags & FLAGS_CBC) && req->info) | ||
573 | memcpy(dev->iv_base, req->info, AES_KEYSIZE_128); | ||
574 | |||
575 | /* assign new context to device */ | ||
576 | ctx->dev = dev; | ||
577 | dev->ctx = ctx; | ||
578 | |||
579 | ret = sahara_hw_descriptor_create(dev); | ||
580 | if (ret < 0) { | ||
581 | spin_lock(&dev->lock); | ||
582 | clear_bit(FLAGS_BUSY, &dev->flags); | ||
583 | spin_unlock(&dev->lock); | ||
584 | dev->req->base.complete(&dev->req->base, ret); | ||
585 | } | ||
586 | } | ||
587 | |||
588 | static int sahara_aes_setkey(struct crypto_ablkcipher *tfm, const u8 *key, | ||
589 | unsigned int keylen) | ||
590 | { | ||
591 | struct sahara_ctx *ctx = crypto_ablkcipher_ctx(tfm); | ||
592 | int ret; | ||
593 | |||
594 | ctx->keylen = keylen; | ||
595 | |||
596 | /* SAHARA only supports 128bit keys */ | ||
597 | if (keylen == AES_KEYSIZE_128) { | ||
598 | memcpy(ctx->key, key, keylen); | ||
599 | ctx->flags |= FLAGS_NEW_KEY; | ||
600 | return 0; | ||
601 | } | ||
602 | |||
603 | if (keylen != AES_KEYSIZE_128 && | ||
604 | keylen != AES_KEYSIZE_192 && keylen != AES_KEYSIZE_256) | ||
605 | return -EINVAL; | ||
606 | |||
607 | /* | ||
608 | * The requested key size is not supported by HW, do a fallback. | ||
609 | */ | ||
610 | ctx->fallback->base.crt_flags &= ~CRYPTO_TFM_REQ_MASK; | ||
611 | ctx->fallback->base.crt_flags |= | ||
612 | (tfm->base.crt_flags & CRYPTO_TFM_REQ_MASK); | ||
613 | |||
614 | ret = crypto_ablkcipher_setkey(ctx->fallback, key, keylen); | ||
615 | if (ret) { | ||
616 | struct crypto_tfm *tfm_aux = crypto_ablkcipher_tfm(tfm); | ||
617 | |||
618 | tfm_aux->crt_flags &= ~CRYPTO_TFM_RES_MASK; | ||
619 | tfm_aux->crt_flags |= | ||
620 | (ctx->fallback->base.crt_flags & CRYPTO_TFM_RES_MASK); | ||
621 | } | ||
622 | return ret; | ||
623 | } | ||
624 | |||
625 | static int sahara_aes_crypt(struct ablkcipher_request *req, unsigned long mode) | ||
626 | { | ||
627 | struct sahara_ctx *ctx = crypto_ablkcipher_ctx( | ||
628 | crypto_ablkcipher_reqtfm(req)); | ||
629 | struct sahara_aes_reqctx *rctx = ablkcipher_request_ctx(req); | ||
630 | struct sahara_dev *dev = dev_ptr; | ||
631 | int err = 0; | ||
632 | int busy; | ||
633 | |||
634 | dev_dbg(dev->device, "nbytes: %d, enc: %d, cbc: %d\n", | ||
635 | req->nbytes, !!(mode & FLAGS_ENCRYPT), !!(mode & FLAGS_CBC)); | ||
636 | |||
637 | if (!IS_ALIGNED(req->nbytes, AES_BLOCK_SIZE)) { | ||
638 | dev_err(dev->device, | ||
639 | "request size is not exact amount of AES blocks\n"); | ||
640 | return -EINVAL; | ||
641 | } | ||
642 | |||
643 | ctx->dev = dev; | ||
644 | |||
645 | rctx->mode = mode; | ||
646 | spin_lock_bh(&dev->lock); | ||
647 | err = ablkcipher_enqueue_request(&dev->queue, req); | ||
648 | busy = test_and_set_bit(FLAGS_BUSY, &dev->flags); | ||
649 | spin_unlock_bh(&dev->lock); | ||
650 | |||
651 | if (!busy) | ||
652 | tasklet_schedule(&dev->queue_task); | ||
653 | |||
654 | return err; | ||
655 | } | ||
656 | |||
657 | static int sahara_aes_ecb_encrypt(struct ablkcipher_request *req) | ||
658 | { | ||
659 | struct crypto_tfm *tfm = | ||
660 | crypto_ablkcipher_tfm(crypto_ablkcipher_reqtfm(req)); | ||
661 | struct sahara_ctx *ctx = crypto_ablkcipher_ctx( | ||
662 | crypto_ablkcipher_reqtfm(req)); | ||
663 | int err; | ||
664 | |||
665 | if (unlikely(ctx->keylen != AES_KEYSIZE_128)) { | ||
666 | ablkcipher_request_set_tfm(req, ctx->fallback); | ||
667 | err = crypto_ablkcipher_encrypt(req); | ||
668 | ablkcipher_request_set_tfm(req, __crypto_ablkcipher_cast(tfm)); | ||
669 | return err; | ||
670 | } | ||
671 | |||
672 | return sahara_aes_crypt(req, FLAGS_ENCRYPT); | ||
673 | } | ||
674 | |||
675 | static int sahara_aes_ecb_decrypt(struct ablkcipher_request *req) | ||
676 | { | ||
677 | struct crypto_tfm *tfm = | ||
678 | crypto_ablkcipher_tfm(crypto_ablkcipher_reqtfm(req)); | ||
679 | struct sahara_ctx *ctx = crypto_ablkcipher_ctx( | ||
680 | crypto_ablkcipher_reqtfm(req)); | ||
681 | int err; | ||
682 | |||
683 | if (unlikely(ctx->keylen != AES_KEYSIZE_128)) { | ||
684 | ablkcipher_request_set_tfm(req, ctx->fallback); | ||
685 | err = crypto_ablkcipher_decrypt(req); | ||
686 | ablkcipher_request_set_tfm(req, __crypto_ablkcipher_cast(tfm)); | ||
687 | return err; | ||
688 | } | ||
689 | |||
690 | return sahara_aes_crypt(req, 0); | ||
691 | } | ||
692 | |||
693 | static int sahara_aes_cbc_encrypt(struct ablkcipher_request *req) | ||
694 | { | ||
695 | struct crypto_tfm *tfm = | ||
696 | crypto_ablkcipher_tfm(crypto_ablkcipher_reqtfm(req)); | ||
697 | struct sahara_ctx *ctx = crypto_ablkcipher_ctx( | ||
698 | crypto_ablkcipher_reqtfm(req)); | ||
699 | int err; | ||
700 | |||
701 | if (unlikely(ctx->keylen != AES_KEYSIZE_128)) { | ||
702 | ablkcipher_request_set_tfm(req, ctx->fallback); | ||
703 | err = crypto_ablkcipher_encrypt(req); | ||
704 | ablkcipher_request_set_tfm(req, __crypto_ablkcipher_cast(tfm)); | ||
705 | return err; | ||
706 | } | ||
707 | |||
708 | return sahara_aes_crypt(req, FLAGS_ENCRYPT | FLAGS_CBC); | ||
709 | } | ||
710 | |||
711 | static int sahara_aes_cbc_decrypt(struct ablkcipher_request *req) | ||
712 | { | ||
713 | struct crypto_tfm *tfm = | ||
714 | crypto_ablkcipher_tfm(crypto_ablkcipher_reqtfm(req)); | ||
715 | struct sahara_ctx *ctx = crypto_ablkcipher_ctx( | ||
716 | crypto_ablkcipher_reqtfm(req)); | ||
717 | int err; | ||
718 | |||
719 | if (unlikely(ctx->keylen != AES_KEYSIZE_128)) { | ||
720 | ablkcipher_request_set_tfm(req, ctx->fallback); | ||
721 | err = crypto_ablkcipher_decrypt(req); | ||
722 | ablkcipher_request_set_tfm(req, __crypto_ablkcipher_cast(tfm)); | ||
723 | return err; | ||
724 | } | ||
725 | |||
726 | return sahara_aes_crypt(req, FLAGS_CBC); | ||
727 | } | ||
728 | |||
729 | static int sahara_aes_cra_init(struct crypto_tfm *tfm) | ||
730 | { | ||
731 | const char *name = tfm->__crt_alg->cra_name; | ||
732 | struct sahara_ctx *ctx = crypto_tfm_ctx(tfm); | ||
733 | |||
734 | ctx->fallback = crypto_alloc_ablkcipher(name, 0, | ||
735 | CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK); | ||
736 | if (IS_ERR(ctx->fallback)) { | ||
737 | pr_err("Error allocating fallback algo %s\n", name); | ||
738 | return PTR_ERR(ctx->fallback); | ||
739 | } | ||
740 | |||
741 | tfm->crt_ablkcipher.reqsize = sizeof(struct sahara_aes_reqctx); | ||
742 | |||
743 | return 0; | ||
744 | } | ||
745 | |||
746 | static void sahara_aes_cra_exit(struct crypto_tfm *tfm) | ||
747 | { | ||
748 | struct sahara_ctx *ctx = crypto_tfm_ctx(tfm); | ||
749 | |||
750 | if (ctx->fallback) | ||
751 | crypto_free_ablkcipher(ctx->fallback); | ||
752 | ctx->fallback = NULL; | ||
753 | } | ||
754 | |||
755 | static struct crypto_alg aes_algs[] = { | ||
756 | { | ||
757 | .cra_name = "ecb(aes)", | ||
758 | .cra_driver_name = "sahara-ecb-aes", | ||
759 | .cra_priority = 300, | ||
760 | .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | | ||
761 | CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK, | ||
762 | .cra_blocksize = AES_BLOCK_SIZE, | ||
763 | .cra_ctxsize = sizeof(struct sahara_ctx), | ||
764 | .cra_alignmask = 0x0, | ||
765 | .cra_type = &crypto_ablkcipher_type, | ||
766 | .cra_module = THIS_MODULE, | ||
767 | .cra_init = sahara_aes_cra_init, | ||
768 | .cra_exit = sahara_aes_cra_exit, | ||
769 | .cra_u.ablkcipher = { | ||
770 | .min_keysize = AES_MIN_KEY_SIZE , | ||
771 | .max_keysize = AES_MAX_KEY_SIZE, | ||
772 | .setkey = sahara_aes_setkey, | ||
773 | .encrypt = sahara_aes_ecb_encrypt, | ||
774 | .decrypt = sahara_aes_ecb_decrypt, | ||
775 | } | ||
776 | }, { | ||
777 | .cra_name = "cbc(aes)", | ||
778 | .cra_driver_name = "sahara-cbc-aes", | ||
779 | .cra_priority = 300, | ||
780 | .cra_flags = CRYPTO_ALG_TYPE_ABLKCIPHER | | ||
781 | CRYPTO_ALG_ASYNC | CRYPTO_ALG_NEED_FALLBACK, | ||
782 | .cra_blocksize = AES_BLOCK_SIZE, | ||
783 | .cra_ctxsize = sizeof(struct sahara_ctx), | ||
784 | .cra_alignmask = 0x0, | ||
785 | .cra_type = &crypto_ablkcipher_type, | ||
786 | .cra_module = THIS_MODULE, | ||
787 | .cra_init = sahara_aes_cra_init, | ||
788 | .cra_exit = sahara_aes_cra_exit, | ||
789 | .cra_u.ablkcipher = { | ||
790 | .min_keysize = AES_MIN_KEY_SIZE , | ||
791 | .max_keysize = AES_MAX_KEY_SIZE, | ||
792 | .ivsize = AES_BLOCK_SIZE, | ||
793 | .setkey = sahara_aes_setkey, | ||
794 | .encrypt = sahara_aes_cbc_encrypt, | ||
795 | .decrypt = sahara_aes_cbc_decrypt, | ||
796 | } | ||
797 | } | ||
798 | }; | ||
799 | |||
800 | static irqreturn_t sahara_irq_handler(int irq, void *data) | ||
801 | { | ||
802 | struct sahara_dev *dev = (struct sahara_dev *)data; | ||
803 | unsigned int stat = sahara_read(dev, SAHARA_REG_STATUS); | ||
804 | unsigned int err = sahara_read(dev, SAHARA_REG_ERRSTATUS); | ||
805 | |||
806 | del_timer(&dev->watchdog); | ||
807 | |||
808 | sahara_write(dev, SAHARA_CMD_CLEAR_INT | SAHARA_CMD_CLEAR_ERR, | ||
809 | SAHARA_REG_CMD); | ||
810 | |||
811 | sahara_decode_status(dev, stat); | ||
812 | |||
813 | if (SAHARA_STATUS_GET_STATE(stat) == SAHARA_STATE_BUSY) { | ||
814 | return IRQ_NONE; | ||
815 | } else if (SAHARA_STATUS_GET_STATE(stat) == SAHARA_STATE_COMPLETE) { | ||
816 | dev->error = 0; | ||
817 | } else { | ||
818 | sahara_decode_error(dev, err); | ||
819 | dev->error = -EINVAL; | ||
820 | } | ||
821 | |||
822 | tasklet_schedule(&dev->done_task); | ||
823 | |||
824 | return IRQ_HANDLED; | ||
825 | } | ||
826 | |||
827 | |||
828 | static int sahara_register_algs(struct sahara_dev *dev) | ||
829 | { | ||
830 | int err, i, j; | ||
831 | |||
832 | for (i = 0; i < ARRAY_SIZE(aes_algs); i++) { | ||
833 | INIT_LIST_HEAD(&aes_algs[i].cra_list); | ||
834 | err = crypto_register_alg(&aes_algs[i]); | ||
835 | if (err) | ||
836 | goto err_aes_algs; | ||
837 | } | ||
838 | |||
839 | return 0; | ||
840 | |||
841 | err_aes_algs: | ||
842 | for (j = 0; j < i; j++) | ||
843 | crypto_unregister_alg(&aes_algs[j]); | ||
844 | |||
845 | return err; | ||
846 | } | ||
847 | |||
848 | static void sahara_unregister_algs(struct sahara_dev *dev) | ||
849 | { | ||
850 | int i; | ||
851 | |||
852 | for (i = 0; i < ARRAY_SIZE(aes_algs); i++) | ||
853 | crypto_unregister_alg(&aes_algs[i]); | ||
854 | } | ||
855 | |||
856 | static struct platform_device_id sahara_platform_ids[] = { | ||
857 | { .name = "sahara-imx27" }, | ||
858 | { /* sentinel */ } | ||
859 | }; | ||
860 | MODULE_DEVICE_TABLE(platform, sahara_platform_ids); | ||
861 | |||
862 | static struct of_device_id sahara_dt_ids[] = { | ||
863 | { .compatible = "fsl,imx27-sahara" }, | ||
864 | { /* sentinel */ } | ||
865 | }; | ||
866 | MODULE_DEVICE_TABLE(platform, sahara_dt_ids); | ||
867 | |||
868 | static int sahara_probe(struct platform_device *pdev) | ||
869 | { | ||
870 | struct sahara_dev *dev; | ||
871 | struct resource *res; | ||
872 | u32 version; | ||
873 | int irq; | ||
874 | int err; | ||
875 | int i; | ||
876 | |||
877 | dev = devm_kzalloc(&pdev->dev, sizeof(struct sahara_dev), GFP_KERNEL); | ||
878 | if (dev == NULL) { | ||
879 | dev_err(&pdev->dev, "unable to alloc data struct.\n"); | ||
880 | return -ENOMEM; | ||
881 | } | ||
882 | |||
883 | dev->device = &pdev->dev; | ||
884 | platform_set_drvdata(pdev, dev); | ||
885 | |||
886 | /* Get the base address */ | ||
887 | res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | ||
888 | if (!res) { | ||
889 | dev_err(&pdev->dev, "failed to get memory region resource\n"); | ||
890 | return -ENODEV; | ||
891 | } | ||
892 | |||
893 | if (devm_request_mem_region(&pdev->dev, res->start, | ||
894 | resource_size(res), SAHARA_NAME) == NULL) { | ||
895 | dev_err(&pdev->dev, "failed to request memory region\n"); | ||
896 | return -ENOENT; | ||
897 | } | ||
898 | dev->regs_base = devm_ioremap(&pdev->dev, res->start, | ||
899 | resource_size(res)); | ||
900 | if (!dev->regs_base) { | ||
901 | dev_err(&pdev->dev, "failed to ioremap address region\n"); | ||
902 | return -ENOENT; | ||
903 | } | ||
904 | |||
905 | /* Get the IRQ */ | ||
906 | irq = platform_get_irq(pdev, 0); | ||
907 | if (irq < 0) { | ||
908 | dev_err(&pdev->dev, "failed to get irq resource\n"); | ||
909 | return irq; | ||
910 | } | ||
911 | |||
912 | if (devm_request_irq(&pdev->dev, irq, sahara_irq_handler, | ||
913 | 0, SAHARA_NAME, dev) < 0) { | ||
914 | dev_err(&pdev->dev, "failed to request irq\n"); | ||
915 | return -ENOENT; | ||
916 | } | ||
917 | |||
918 | /* clocks */ | ||
919 | dev->clk_ipg = devm_clk_get(&pdev->dev, "ipg"); | ||
920 | if (IS_ERR(dev->clk_ipg)) { | ||
921 | dev_err(&pdev->dev, "Could not get ipg clock\n"); | ||
922 | return PTR_ERR(dev->clk_ipg); | ||
923 | } | ||
924 | |||
925 | dev->clk_ahb = devm_clk_get(&pdev->dev, "ahb"); | ||
926 | if (IS_ERR(dev->clk_ahb)) { | ||
927 | dev_err(&pdev->dev, "Could not get ahb clock\n"); | ||
928 | return PTR_ERR(dev->clk_ahb); | ||
929 | } | ||
930 | |||
931 | /* Allocate HW descriptors */ | ||
932 | dev->hw_desc[0] = dma_alloc_coherent(&pdev->dev, | ||
933 | SAHARA_MAX_HW_DESC * sizeof(struct sahara_hw_desc), | ||
934 | &dev->hw_phys_desc[0], GFP_KERNEL); | ||
935 | if (!dev->hw_desc[0]) { | ||
936 | dev_err(&pdev->dev, "Could not allocate hw descriptors\n"); | ||
937 | return -ENOMEM; | ||
938 | } | ||
939 | dev->hw_desc[1] = dev->hw_desc[0] + 1; | ||
940 | dev->hw_phys_desc[1] = dev->hw_phys_desc[0] + | ||
941 | sizeof(struct sahara_hw_desc); | ||
942 | |||
943 | /* Allocate space for iv and key */ | ||
944 | dev->key_base = dma_alloc_coherent(&pdev->dev, 2 * AES_KEYSIZE_128, | ||
945 | &dev->key_phys_base, GFP_KERNEL); | ||
946 | if (!dev->key_base) { | ||
947 | dev_err(&pdev->dev, "Could not allocate memory for key\n"); | ||
948 | err = -ENOMEM; | ||
949 | goto err_key; | ||
950 | } | ||
951 | dev->iv_base = dev->key_base + AES_KEYSIZE_128; | ||
952 | dev->iv_phys_base = dev->key_phys_base + AES_KEYSIZE_128; | ||
953 | |||
954 | /* Allocate space for HW links */ | ||
955 | dev->hw_link[0] = dma_alloc_coherent(&pdev->dev, | ||
956 | SAHARA_MAX_HW_LINK * sizeof(struct sahara_hw_link), | ||
957 | &dev->hw_phys_link[0], GFP_KERNEL); | ||
958 | if (!dev->hw_link) { | ||
959 | dev_err(&pdev->dev, "Could not allocate hw links\n"); | ||
960 | err = -ENOMEM; | ||
961 | goto err_link; | ||
962 | } | ||
963 | for (i = 1; i < SAHARA_MAX_HW_LINK; i++) { | ||
964 | dev->hw_phys_link[i] = dev->hw_phys_link[i - 1] + | ||
965 | sizeof(struct sahara_hw_link); | ||
966 | dev->hw_link[i] = dev->hw_link[i - 1] + 1; | ||
967 | } | ||
968 | |||
969 | crypto_init_queue(&dev->queue, SAHARA_QUEUE_LENGTH); | ||
970 | |||
971 | dev_ptr = dev; | ||
972 | |||
973 | tasklet_init(&dev->queue_task, sahara_aes_queue_task, | ||
974 | (unsigned long)dev); | ||
975 | tasklet_init(&dev->done_task, sahara_aes_done_task, | ||
976 | (unsigned long)dev); | ||
977 | |||
978 | init_timer(&dev->watchdog); | ||
979 | dev->watchdog.function = &sahara_watchdog; | ||
980 | dev->watchdog.data = (unsigned long)dev; | ||
981 | |||
982 | clk_prepare_enable(dev->clk_ipg); | ||
983 | clk_prepare_enable(dev->clk_ahb); | ||
984 | |||
985 | version = sahara_read(dev, SAHARA_REG_VERSION); | ||
986 | if (version != SAHARA_VERSION_3) { | ||
987 | dev_err(&pdev->dev, "SAHARA version %d not supported\n", | ||
988 | version); | ||
989 | err = -ENODEV; | ||
990 | goto err_algs; | ||
991 | } | ||
992 | |||
993 | sahara_write(dev, SAHARA_CMD_RESET | SAHARA_CMD_MODE_BATCH, | ||
994 | SAHARA_REG_CMD); | ||
995 | sahara_write(dev, SAHARA_CONTROL_SET_THROTTLE(0) | | ||
996 | SAHARA_CONTROL_SET_MAXBURST(8) | | ||
997 | SAHARA_CONTROL_RNG_AUTORSD | | ||
998 | SAHARA_CONTROL_ENABLE_INT, | ||
999 | SAHARA_REG_CONTROL); | ||
1000 | |||
1001 | err = sahara_register_algs(dev); | ||
1002 | if (err) | ||
1003 | goto err_algs; | ||
1004 | |||
1005 | dev_info(&pdev->dev, "SAHARA version %d initialized\n", version); | ||
1006 | |||
1007 | return 0; | ||
1008 | |||
1009 | err_algs: | ||
1010 | dma_free_coherent(&pdev->dev, | ||
1011 | SAHARA_MAX_HW_LINK * sizeof(struct sahara_hw_link), | ||
1012 | dev->hw_link[0], dev->hw_phys_link[0]); | ||
1013 | clk_disable_unprepare(dev->clk_ipg); | ||
1014 | clk_disable_unprepare(dev->clk_ahb); | ||
1015 | dev_ptr = NULL; | ||
1016 | err_link: | ||
1017 | dma_free_coherent(&pdev->dev, | ||
1018 | 2 * AES_KEYSIZE_128, | ||
1019 | dev->key_base, dev->key_phys_base); | ||
1020 | err_key: | ||
1021 | dma_free_coherent(&pdev->dev, | ||
1022 | SAHARA_MAX_HW_DESC * sizeof(struct sahara_hw_desc), | ||
1023 | dev->hw_desc[0], dev->hw_phys_desc[0]); | ||
1024 | |||
1025 | return err; | ||
1026 | } | ||
1027 | |||
1028 | static int sahara_remove(struct platform_device *pdev) | ||
1029 | { | ||
1030 | struct sahara_dev *dev = platform_get_drvdata(pdev); | ||
1031 | |||
1032 | dma_free_coherent(&pdev->dev, | ||
1033 | SAHARA_MAX_HW_LINK * sizeof(struct sahara_hw_link), | ||
1034 | dev->hw_link[0], dev->hw_phys_link[0]); | ||
1035 | dma_free_coherent(&pdev->dev, | ||
1036 | 2 * AES_KEYSIZE_128, | ||
1037 | dev->key_base, dev->key_phys_base); | ||
1038 | dma_free_coherent(&pdev->dev, | ||
1039 | SAHARA_MAX_HW_DESC * sizeof(struct sahara_hw_desc), | ||
1040 | dev->hw_desc[0], dev->hw_phys_desc[0]); | ||
1041 | |||
1042 | tasklet_kill(&dev->done_task); | ||
1043 | tasklet_kill(&dev->queue_task); | ||
1044 | |||
1045 | sahara_unregister_algs(dev); | ||
1046 | |||
1047 | clk_disable_unprepare(dev->clk_ipg); | ||
1048 | clk_disable_unprepare(dev->clk_ahb); | ||
1049 | |||
1050 | dev_ptr = NULL; | ||
1051 | |||
1052 | return 0; | ||
1053 | } | ||
1054 | |||
1055 | static struct platform_driver sahara_driver = { | ||
1056 | .probe = sahara_probe, | ||
1057 | .remove = sahara_remove, | ||
1058 | .driver = { | ||
1059 | .name = SAHARA_NAME, | ||
1060 | .owner = THIS_MODULE, | ||
1061 | .of_match_table = of_match_ptr(sahara_dt_ids), | ||
1062 | }, | ||
1063 | .id_table = sahara_platform_ids, | ||
1064 | }; | ||
1065 | |||
1066 | module_platform_driver(sahara_driver); | ||
1067 | |||
1068 | MODULE_LICENSE("GPL"); | ||
1069 | MODULE_AUTHOR("Javier Martin <javier.martin@vista-silicon.com>"); | ||
1070 | MODULE_DESCRIPTION("SAHARA2 HW crypto accelerator"); | ||