udhcpc: Pass all events to udhcpc.user
[openwrt/svn-archive/archive.git] / target / linux / s3c24xx / files-2.6.30 / drivers / mfd / glamo / glamo-mci.c
1 /*
2 * linux/drivers/mmc/host/glamo-mmc.c - Glamo MMC driver
3 *
4 * Copyright (C) 2007 Openmoko, Inc, Andy Green <andy@openmoko.com>
5 * Based on S3C MMC driver that was:
6 * Copyright (C) 2004-2006 maintech GmbH, Thomas Kleffel <tk@maintech.de>
7 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License version 2 as
10 * published by the Free Software Foundation.
11 */
12
13 #include <linux/module.h>
14 #include <linux/dma-mapping.h>
15 #include <linux/clk.h>
16 #include <linux/mmc/mmc.h>
17 #include <linux/mmc/sd.h>
18 #include <linux/mmc/host.h>
19 #include <linux/platform_device.h>
20 #include <linux/irq.h>
21 #include <linux/delay.h>
22 #include <linux/interrupt.h>
23 #include <linux/spinlock.h>
24 #include <linux/workqueue.h>
25 #include <linux/crc7.h>
26
27 #include <asm/dma.h>
28 #include <asm/dma-mapping.h>
29 #include <asm/io.h>
30
31 #include "glamo-mci.h"
32 #include "glamo-core.h"
33 #include "glamo-regs.h"
34
35 #define DRIVER_NAME "glamo-mci"
36
37 static void glamo_mci_send_request(struct mmc_host *mmc);
38 static void glamo_mci_send_command(struct glamo_mci_host *host,
39 struct mmc_command *cmd);
40
41 /*
42 * Max SD clock rate
43 *
44 * held at /(3 + 1) due to concerns of 100R recommended series resistor
45 * allows 16MHz @ 4-bit --> 8MBytes/sec raw
46 *
47 * you can override this on kernel commandline using
48 *
49 * glamo_mci.sd_max_clk=10000000
50 *
51 * for example
52 */
53
54 static int sd_max_clk = 50000000 / 3;
55 module_param(sd_max_clk, int, 0644);
56
57 /*
58 * Slow SD clock rate
59 *
60 * you can override this on kernel commandline using
61 *
62 * glamo_mci.sd_slow_ratio=8
63 *
64 * for example
65 *
66 * platform callback is used to decide effective clock rate, if not
67 * defined then max is used, if defined and returns nonzero, rate is
68 * divided by this factor
69 */
70
71 static int sd_slow_ratio = 8;
72 module_param(sd_slow_ratio, int, 0644);
73
74 /*
75 * Post-power SD clock rate
76 *
77 * you can override this on kernel commandline using
78 *
79 * glamo_mci.sd_post_power_clock=1000000
80 *
81 * for example
82 *
83 * After changing power to card, clock is held at this rate until first bulk
84 * transfer completes
85 */
86
87 static int sd_post_power_clock = 1000000;
88 module_param(sd_post_power_clock, int, 0644);
89
90
91 /*
92 * SD Signal drive strength
93 *
94 * you can override this on kernel commandline using
95 *
96 * glamo_mci.sd_drive=0
97 *
98 * for example
99 */
100
101 static int sd_drive;
102 module_param(sd_drive, int, 0644);
103
104 /*
105 * SD allow SD clock to run while idle
106 *
107 * you can override this on kernel commandline using
108 *
109 * glamo_mci.sd_idleclk=0
110 *
111 * for example
112 */
113
114 static int sd_idleclk = 0; /* disallow idle clock by default */
115 module_param(sd_idleclk, int, 0644);
116
117 /* used to stash real idleclk state in suspend: we force it to run in there */
118 static int suspend_sd_idleclk;
119
120 static inline void glamo_reg_write(struct glamo_mci_host *glamo,
121 u_int16_t reg, u_int16_t val)
122 {
123 writew(val, glamo->mmio_base + reg);
124 }
125
126 static inline u_int16_t glamo_reg_read(struct glamo_mci_host *glamo,
127 u_int16_t reg)
128 {
129 return readw(glamo->mmio_base + reg);
130 }
131
132 static void glamo_reg_set_bit_mask(struct glamo_mci_host *glamo,
133 u_int16_t reg, u_int16_t mask,
134 u_int16_t val)
135 {
136 u_int16_t tmp;
137
138 val &= mask;
139
140 tmp = glamo_reg_read(glamo, reg);
141 tmp &= ~mask;
142 tmp |= val;
143 glamo_reg_write(glamo, reg, tmp);
144 }
145
146 static void do_pio_read(struct glamo_mci_host *host)
147 {
148 struct scatterlist *sg;
149 u16 __iomem *from_ptr = host->data_base;
150 struct mmc_data *data = host->mrq->data;
151 void *sg_pointer;
152
153 dev_dbg(&host->pdev->dev, "pio_read():\n");
154 for (sg = data->sg; sg; sg = sg_next(sg)) {
155 sg_pointer = page_address(sg_page(sg)) + sg->offset;
156
157
158 memcpy(sg_pointer, from_ptr, sg->length);
159 from_ptr += sg->length >> 1;
160
161 data->bytes_xfered += sg->length;
162 }
163
164 dev_dbg(&host->pdev->dev, "pio_read(): "
165 "complete (no more data).\n");
166 }
167
168 static void do_pio_write(struct glamo_mci_host *host)
169 {
170 struct scatterlist *sg;
171 u16 __iomem *to_ptr = host->data_base;
172 struct mmc_data *data = host->mrq->data;
173 void *sg_pointer;
174
175 dev_dbg(&host->pdev->dev, "pio_write():\n");
176 for (sg = data->sg; sg; sg = sg_next(sg)) {
177 sg_pointer = page_address(sg_page(sg)) + sg->offset;
178
179 data->bytes_xfered += sg->length;
180
181 memcpy(to_ptr, sg_pointer, sg->length);
182 to_ptr += sg->length >> 1;
183 }
184
185 dev_dbg(&host->pdev->dev, "pio_write(): complete\n");
186 }
187
188 static void glamo_mci_fix_card_div(struct glamo_mci_host *host, int div)
189 {
190 unsigned long flags;
191
192 spin_lock_irqsave(&host->pdata->pglamo->lock, flags);
193
194 if (div < 0) {
195 /* stop clock - remove clock from divider input */
196 writew(readw(host->pdata->pglamo->base +
197 GLAMO_REG_CLOCK_GEN5_1) & (~GLAMO_CLOCK_GEN51_EN_DIV_TCLK),
198 host->pdata->pglamo->base + GLAMO_REG_CLOCK_GEN5_1);
199 } else {
200
201 if (host->force_slow_during_powerup)
202 div = host->clk_rate / sd_post_power_clock;
203 else if (host->pdata->glamo_mci_use_slow &&
204 host->pdata->glamo_mci_use_slow())
205 div = div * sd_slow_ratio;
206
207 if (div > 255)
208 div = 255;
209 /*
210 * set the nearest prescaler factor
211 *
212 * register shared with SCLK divisor -- no chance of race because
213 * we don't use sensor interface
214 */
215 writew((readw(host->pdata->pglamo->base +
216 GLAMO_REG_CLOCK_GEN8) & 0xff00) | div,
217 host->pdata->pglamo->base + GLAMO_REG_CLOCK_GEN8);
218 /* enable clock to divider input */
219 writew(readw(host->pdata->pglamo->base +
220 GLAMO_REG_CLOCK_GEN5_1) | GLAMO_CLOCK_GEN51_EN_DIV_TCLK,
221 host->pdata->pglamo->base + GLAMO_REG_CLOCK_GEN5_1);
222 mdelay(5);
223 }
224 spin_unlock_irqrestore(&host->pdata->pglamo->lock, flags);
225 }
226
227 static int glamo_mci_set_card_clock(struct glamo_mci_host *host, int freq)
228 {
229 int div = 0;
230 int real_rate = 0;
231
232 if (freq) {
233 /* Set clock */
234 for (div = 0; div < 255; div++) {
235 real_rate = host->clk_rate / (div + 1);
236 if (real_rate <= freq)
237 break;
238 }
239
240 host->clk_div = div;
241 glamo_mci_fix_card_div(host, div);
242 } else {
243 /* stop clock */
244 host->clk_div = 0xff;
245
246 if (!sd_idleclk && !host->force_slow_during_powerup)
247 /* clock off */
248 glamo_mci_fix_card_div(host, -1);
249 }
250 host->real_rate = real_rate;
251 return real_rate;
252 }
253
254
255 static void glamo_mci_irq_worker(struct work_struct *work)
256 {
257 struct glamo_mci_host *host =
258 container_of(work, struct glamo_mci_host, irq_work);
259 struct mmc_command *cmd = host->mrq->cmd;
260
261 if (cmd->data->flags & MMC_DATA_READ) {
262 do_pio_read(host);
263 }
264
265 /* issue STOP if we have been given one to use */
266 if (host->mrq->stop) {
267 glamo_mci_send_command(host, host->mrq->stop);
268 }
269
270 if (!sd_idleclk && !host->force_slow_during_powerup)
271 /* clock off */
272 glamo_mci_fix_card_div(host, -1);
273
274 host->mrq = NULL;
275 mmc_request_done(host->mmc, cmd->mrq);
276 }
277
278 static irqreturn_t glamo_mci_irq(int irq, void *devid)
279 {
280 struct glamo_mci_host *host = (struct glamo_mci_host*)devid;
281 u16 status;
282 struct mmc_command *cmd;
283 unsigned long flags;
284
285 if (host->suspending) { /* bad news, dangerous time */
286 dev_err(&host->pdev->dev, "****glamo_mci_irq before resumed\n");
287 goto leave;
288 }
289
290 if (!host->mrq)
291 goto leave;
292 cmd = host->mrq->cmd;
293 if (!cmd)
294 goto leave;
295
296 spin_lock_irqsave(&host->lock, flags);
297
298 status = readw(host->mmio_base + GLAMO_REG_MMC_RB_STAT1);
299 dev_dbg(&host->pdev->dev, "status = 0x%04x\n", status);
300
301 /* we ignore a data timeout report if we are also told the data came */
302 if (status & GLAMO_STAT1_MMC_RB_DRDY)
303 status &= ~GLAMO_STAT1_MMC_DTOUT;
304
305 if (status & (GLAMO_STAT1_MMC_RTOUT |
306 GLAMO_STAT1_MMC_DTOUT))
307 cmd->error = -ETIMEDOUT;
308 if (status & (GLAMO_STAT1_MMC_BWERR |
309 GLAMO_STAT1_MMC_BRERR))
310 cmd->error = -EILSEQ;
311 if (cmd->error) {
312 dev_info(&host->pdev->dev, "Error after cmd: 0x%x\n", status);
313 goto done;
314 }
315
316 /*
317 * disable the initial slow start after first bulk transfer
318 */
319 if (host->force_slow_during_powerup)
320 host->force_slow_during_powerup--;
321
322 /*
323 * we perform the memcpy out of Glamo memory outside of IRQ context
324 * so we don't block other interrupts
325 */
326 schedule_work(&host->irq_work);
327
328 goto unlock;
329
330 done:
331 host->mrq = NULL;
332 mmc_request_done(host->mmc, cmd->mrq);
333 unlock:
334 spin_unlock_irqrestore(&host->lock, flags);
335 leave:
336 return IRQ_HANDLED;
337 }
338
339 static void glamo_mci_send_command(struct glamo_mci_host *host,
340 struct mmc_command *cmd)
341 {
342 u8 u8a[6];
343 u16 fire = 0;
344 unsigned int timeout = 1000000;
345 u16 * reg_resp = (u16 *)(host->mmio_base + GLAMO_REG_MMC_CMD_RSP1);
346 u16 status;
347
348 /* if we can't do it, reject as busy */
349 if (!readw(host->mmio_base + GLAMO_REG_MMC_RB_STAT1) &
350 GLAMO_STAT1_MMC_IDLE) {
351 host->mrq = NULL;
352 cmd->error = -EBUSY;
353 mmc_request_done(host->mmc, host->mrq);
354 return;
355 }
356
357 /* create an array in wire order for CRC computation */
358 u8a[0] = 0x40 | (cmd->opcode & 0x3f);
359 u8a[1] = (u8)(cmd->arg >> 24);
360 u8a[2] = (u8)(cmd->arg >> 16);
361 u8a[3] = (u8)(cmd->arg >> 8);
362 u8a[4] = (u8)cmd->arg;
363 u8a[5] = (crc7(0, u8a, 5) << 1) | 0x01; /* crc7 on first 5 bytes of packet */
364
365 /* issue the wire-order array including CRC in register order */
366 writew((u8a[4] << 8) | u8a[5], host->mmio_base + GLAMO_REG_MMC_CMD_REG1);
367 writew((u8a[2] << 8) | u8a[3], host->mmio_base + GLAMO_REG_MMC_CMD_REG2);
368 writew((u8a[0] << 8) | u8a[1], host->mmio_base + GLAMO_REG_MMC_CMD_REG3);
369
370 /* command index toggle */
371 fire |= (host->request_counter & 1) << 12;
372
373 /* set type of command */
374 switch (mmc_cmd_type(cmd)) {
375 case MMC_CMD_BC:
376 fire |= GLAMO_FIRE_MMC_CMDT_BNR;
377 break;
378 case MMC_CMD_BCR:
379 fire |= GLAMO_FIRE_MMC_CMDT_BR;
380 break;
381 case MMC_CMD_AC:
382 fire |= GLAMO_FIRE_MMC_CMDT_AND;
383 break;
384 case MMC_CMD_ADTC:
385 fire |= GLAMO_FIRE_MMC_CMDT_AD;
386 break;
387 }
388 /*
389 * if it expects a response, set the type expected
390 *
391 * R1, Length : 48bit, Normal response
392 * R1b, Length : 48bit, same R1, but added card busy status
393 * R2, Length : 136bit (really 128 bits with CRC snipped)
394 * R3, Length : 48bit (OCR register value)
395 * R4, Length : 48bit, SDIO_OP_CONDITION, Reverse SDIO Card
396 * R5, Length : 48bit, IO_RW_DIRECTION, Reverse SDIO Card
397 * R6, Length : 48bit (RCA register)
398 * R7, Length : 48bit (interface condition, VHS(voltage supplied),
399 * check pattern, CRC7)
400 */
401 switch (mmc_resp_type(cmd)) {
402 case MMC_RSP_R1: /* same index as R6 and R7 */
403 fire |= GLAMO_FIRE_MMC_RSPT_R1;
404 break;
405 case MMC_RSP_R1B:
406 fire |= GLAMO_FIRE_MMC_RSPT_R1b;
407 break;
408 case MMC_RSP_R2:
409 fire |= GLAMO_FIRE_MMC_RSPT_R2;
410 break;
411 case MMC_RSP_R3:
412 fire |= GLAMO_FIRE_MMC_RSPT_R3;
413 break;
414 /* R4 and R5 supported by chip not defined in linux/mmc/core.h (sdio) */
415 }
416 /*
417 * From the command index, set up the command class in the host ctrllr
418 *
419 * missing guys present on chip but couldn't figure out how to use yet:
420 * 0x0 "stream read"
421 * 0x9 "cancel running command"
422 */
423 switch (cmd->opcode) {
424 case MMC_READ_SINGLE_BLOCK:
425 fire |= GLAMO_FIRE_MMC_CC_SBR; /* single block read */
426 break;
427 case MMC_SWITCH: /* 64 byte payload */
428 case SD_APP_SEND_SCR:
429 case MMC_READ_MULTIPLE_BLOCK:
430 /* we will get an interrupt off this */
431 if (!cmd->mrq->stop)
432 /* multiblock no stop */
433 fire |= GLAMO_FIRE_MMC_CC_MBRNS;
434 else
435 /* multiblock with stop */
436 fire |= GLAMO_FIRE_MMC_CC_MBRS;
437 break;
438 case MMC_WRITE_BLOCK:
439 fire |= GLAMO_FIRE_MMC_CC_SBW; /* single block write */
440 break;
441 case MMC_WRITE_MULTIPLE_BLOCK:
442 if (cmd->mrq->stop)
443 /* multiblock with stop */
444 fire |= GLAMO_FIRE_MMC_CC_MBWS;
445 else
446 /* multiblock NO stop-- 'RESERVED'? */
447 fire |= GLAMO_FIRE_MMC_CC_MBWNS;
448 break;
449 case MMC_STOP_TRANSMISSION:
450 fire |= GLAMO_FIRE_MMC_CC_STOP; /* STOP */
451 break;
452 default:
453 fire |= GLAMO_FIRE_MMC_CC_BASIC; /* "basic command" */
454 break;
455 }
456 /* always largest timeout */
457 writew(0xfff, host->mmio_base + GLAMO_REG_MMC_TIMEOUT);
458
459 /* Generate interrupt on txfer */
460 glamo_reg_set_bit_mask(host, GLAMO_REG_MMC_BASIC, ~0x3e,
461 0x0800 | GLAMO_BASIC_MMC_NO_CLK_RD_WAIT |
462 GLAMO_BASIC_MMC_EN_COMPL_INT | (sd_drive << 6));
463
464 /* send the command out on the wire */
465 /* dev_info(&host->pdev->dev, "Using FIRE %04X\n", fire); */
466 writew(fire, host->mmio_base + GLAMO_REG_MMC_CMD_FIRE);
467
468 /* we are deselecting card? because it isn't going to ack then... */
469 if ((cmd->opcode == 7) && (cmd->arg == 0))
470 return;
471
472 /*
473 * we must spin until response is ready or timed out
474 * -- we don't get interrupts unless there is a bulk rx
475 */
476 udelay(5);
477 do
478 status = readw(host->mmio_base + GLAMO_REG_MMC_RB_STAT1);
479 while (((((status >> 15) & 1) != (host->request_counter & 1)) ||
480 (!(status & (GLAMO_STAT1_MMC_RB_RRDY |
481 GLAMO_STAT1_MMC_RTOUT |
482 GLAMO_STAT1_MMC_DTOUT |
483 GLAMO_STAT1_MMC_BWERR |
484 GLAMO_STAT1_MMC_BRERR)))) && (timeout--));
485
486 if ((status & (GLAMO_STAT1_MMC_RTOUT |
487 GLAMO_STAT1_MMC_DTOUT)) ||
488 (timeout == 0)) {
489 cmd->error = -ETIMEDOUT;
490 } else if (status & (GLAMO_STAT1_MMC_BWERR |
491 GLAMO_STAT1_MMC_BRERR)) {
492 cmd->error = -EILSEQ;
493 }
494
495 if (cmd->flags & MMC_RSP_PRESENT) {
496 if (cmd->flags & MMC_RSP_136) {
497 cmd->resp[3] = readw(&reg_resp[0]) |
498 (readw(&reg_resp[1]) << 16);
499 cmd->resp[2] = readw(&reg_resp[2]) |
500 (readw(&reg_resp[3]) << 16);
501 cmd->resp[1] = readw(&reg_resp[4]) |
502 (readw(&reg_resp[5]) << 16);
503 cmd->resp[0] = readw(&reg_resp[6]) |
504 (readw(&reg_resp[7]) << 16);
505 } else {
506 cmd->resp[0] = (readw(&reg_resp[0]) >> 8) |
507 (readw(&reg_resp[1]) << 8) |
508 ((readw(&reg_resp[2])) << 24);
509 }
510 }
511 }
512
513 static int glamo_mci_prepare_pio(struct glamo_mci_host *host,
514 struct mmc_data *data)
515 {
516 /* set up the block info */
517 writew(data->blksz, host->mmio_base + GLAMO_REG_MMC_DATBLKLEN);
518 writew(data->blocks, host->mmio_base + GLAMO_REG_MMC_DATBLKCNT);
519 dev_dbg(&host->pdev->dev, "(blksz=%d, count=%d)\n",
520 data->blksz, data->blocks);
521 data->bytes_xfered = 0;
522
523 /* if write, prep the write into the shared RAM before the command */
524 if (data->flags & MMC_DATA_WRITE) {
525 do_pio_write(host);
526 }
527 return 0;
528 }
529
530 static void glamo_mci_send_request(struct mmc_host *mmc)
531 {
532 struct glamo_mci_host *host = mmc_priv(mmc);
533 struct mmc_request *mrq = host->mrq;
534 struct mmc_command *cmd = mrq->cmd;
535 int timeout = 1000000;
536
537 host->request_counter++;
538 /* this guy has data to read/write? */
539 if (cmd->data) {
540 if(glamo_mci_prepare_pio(host, cmd->data)) {
541 cmd->data->error = -EIO;
542 goto done;
543 }
544 }
545
546 dev_dbg(&host->pdev->dev,"cmd 0x%x, "
547 "arg 0x%x data=%p mrq->stop=%p flags 0x%x\n",
548 cmd->opcode, cmd->arg, cmd->data, cmd->mrq->stop,
549 cmd->flags);
550
551 /* resume requested clock rate
552 * scale it down by sd_slow_ratio if platform requests it
553 */
554 glamo_mci_fix_card_div(host, host->clk_div);
555
556 glamo_mci_send_command(host, cmd);
557
558 /*
559 * if we don't have bulk data to take care of, we're done
560 */
561 if (!cmd->data || cmd->error)
562 goto done;
563
564 /*
565 * Otherwise can can use the interrupt as async completion --
566 * if there is read data coming, or we wait for write data to complete,
567 * exit without mmc_request_done() as the payload interrupt
568 * will service it
569 */
570 dev_dbg(&host->pdev->dev, "Waiting for payload data\n");
571 /*
572 * if the glamo INT# line isn't wired (*cough* it can happen)
573 * I'm afraid we have to spin on the IRQ status bit and "be
574 * our own INT# line"
575 */
576 if (!host->pdata->pglamo->irq_works) {
577 /*
578 * we have faith we will get an "interrupt"...
579 * but something insane like suspend problems can mean
580 * we spin here forever, so we timeout after a LONG time
581 */
582 while ((!(readw(host->pdata->pglamo->base +
583 GLAMO_REG_IRQ_STATUS) & GLAMO_IRQ_MMC)) &&
584 (timeout--));
585
586 if (timeout < 0) {
587 if (cmd->data->error)
588 cmd->data->error = -ETIMEDOUT;
589 dev_err(&host->pdev->dev, "Payload timeout\n");
590 goto bail;
591 }
592 /* ack this interrupt source */
593 writew(GLAMO_IRQ_MMC, host->pdata->pglamo->base +
594 GLAMO_REG_IRQ_CLEAR);
595
596 /* yay we are an interrupt controller! -- call the ISR
597 * it will stop clock to card
598 */
599 glamo_mci_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), host);
600 }
601 return;
602 done:
603 host->mrq = NULL;
604 mmc_request_done(host->mmc, cmd->mrq);
605 bail:
606 if (!sd_idleclk && !host->force_slow_during_powerup)
607 /* stop the clock to card */
608 glamo_mci_fix_card_div(host, -1);
609 }
610
611 static void glamo_mci_request(struct mmc_host *mmc, struct mmc_request *mrq)
612 {
613 struct glamo_mci_host *host = mmc_priv(mmc);
614
615 host->mrq = mrq;
616 glamo_mci_send_request(mmc);
617 }
618
619 static void glamo_mci_set_ios(struct mmc_host *mmc, struct mmc_ios *ios)
620 {
621 struct glamo_mci_host *host = mmc_priv(mmc);
622 int bus_width = 0;
623 int powering = 0;
624
625 if (host->suspending) {
626 dev_err(&host->pdev->dev, "IGNORING glamo_mci_set_ios while "
627 "suspended\n");
628 return;
629 }
630
631 /* Set power */
632 switch(ios->power_mode) {
633 case MMC_POWER_UP:
634 mmc_regulator_set_ocr(host->regulator, ios->vdd);
635 host->vdd_current = ios->vdd;
636 break;
637 case MMC_POWER_ON:
638 /*
639 * we should use very slow clock until first bulk
640 * transfer completes OK
641 */
642 host->force_slow_during_powerup = 1;
643
644 if (host->vdd_current != ios->vdd) {
645 mmc_regulator_set_ocr(host->regulator, ios->vdd);
646 host->vdd_current = ios->vdd;
647 }
648 if (host->power_mode_current == MMC_POWER_OFF) {
649 glamo_engine_enable(host->pdata->pglamo,
650 GLAMO_ENGINE_MMC);
651 powering = 1;
652 }
653 break;
654
655 case MMC_POWER_OFF:
656 default:
657 if (host->power_mode_current == MMC_POWER_OFF)
658 break;
659 /* never want clocking with dead card */
660 glamo_mci_fix_card_div(host, -1);
661
662 glamo_engine_disable(host->pdata->pglamo,
663 GLAMO_ENGINE_MMC);
664
665 mmc_regulator_set_ocr(host->regulator, 0);
666 host->vdd_current = -1;
667 break;
668 }
669 host->power_mode_current = ios->power_mode;
670
671 glamo_mci_set_card_clock(host, ios->clock);
672
673 /* after power-up, we are meant to give it >= 74 clocks so it can
674 * initialize itself. Doubt any modern cards need it but anyway...
675 */
676 if (powering)
677 mdelay(1);
678
679 if (!sd_idleclk && !host->force_slow_during_powerup)
680 /* stop the clock to card, because we are idle until transfer */
681 glamo_mci_fix_card_div(host, -1);
682
683 if ((ios->power_mode == MMC_POWER_ON) ||
684 (ios->power_mode == MMC_POWER_UP)) {
685 dev_info(&host->pdev->dev,
686 "powered (vdd = %d) clk: %lukHz div=%d (req: %ukHz). "
687 "Bus width=%d\n",(int)ios->vdd,
688 host->real_rate / 1000, (int)host->clk_div,
689 ios->clock / 1000, (int)ios->bus_width);
690 } else
691 dev_info(&host->pdev->dev, "glamo_mci_set_ios: power down.\n");
692
693 /* set bus width */
694 if (ios->bus_width == MMC_BUS_WIDTH_4)
695 bus_width = GLAMO_BASIC_MMC_EN_4BIT_DATA;
696 glamo_reg_set_bit_mask(host, GLAMO_REG_MMC_BASIC,
697 GLAMO_BASIC_MMC_EN_4BIT_DATA |
698 GLAMO_BASIC_MMC_EN_DR_STR0 |
699 GLAMO_BASIC_MMC_EN_DR_STR1,
700 bus_width | sd_drive << 6);
701 }
702
703
704 /*
705 * no physical write protect supported by us
706 */
707 static int glamo_mci_get_ro(struct mmc_host *mmc)
708 {
709 return 0;
710 }
711
712 static struct mmc_host_ops glamo_mci_ops = {
713 .request = glamo_mci_request,
714 .set_ios = glamo_mci_set_ios,
715 .get_ro = glamo_mci_get_ro,
716 };
717
718 static int glamo_mci_probe(struct platform_device *pdev)
719 {
720 struct mmc_host *mmc;
721 struct glamo_mci_host *host;
722 int ret;
723
724 dev_info(&pdev->dev, "glamo_mci driver (C)2007 Openmoko, Inc\n");
725
726 mmc = mmc_alloc_host(sizeof(struct glamo_mci_host), &pdev->dev);
727 if (!mmc) {
728 ret = -ENOMEM;
729 goto probe_out;
730 }
731
732 host = mmc_priv(mmc);
733 host->mmc = mmc;
734 host->pdev = pdev;
735 host->pdata = pdev->dev.platform_data;
736 host->power_mode_current = MMC_POWER_OFF;
737
738 spin_lock_init(&host->lock);
739 INIT_WORK(&host->irq_work, glamo_mci_irq_worker);
740
741 host->regulator = regulator_get(&pdev->dev, "SD_3V3");
742 if (!host->regulator) {
743 dev_err(&pdev->dev, "Cannot proceed without regulator.\n");
744 ret = -ENODEV;
745 goto probe_free_host;
746 }
747
748 host->mmio_mem = platform_get_resource(pdev, IORESOURCE_MEM, 0);
749 if (!host->mmio_mem) {
750 dev_err(&pdev->dev,
751 "failed to get io memory region resouce.\n");
752 ret = -ENOENT;
753 goto probe_regulator_put;
754 }
755
756 host->mmio_mem = request_mem_region(host->mmio_mem->start,
757 resource_size(host->mmio_mem),
758 pdev->name);
759
760 if (!host->mmio_mem) {
761 dev_err(&pdev->dev, "failed to request io memory region.\n");
762 ret = -ENOENT;
763 goto probe_regulator_put;
764 }
765
766 host->mmio_base = ioremap(host->mmio_mem->start,
767 resource_size(host->mmio_mem));
768 if (!host->mmio_base) {
769 dev_err(&pdev->dev, "failed to ioremap() io memory region.\n");
770 ret = -EINVAL;
771 goto probe_free_mem_region_mmio;
772 }
773
774
775 /* Get ahold of our data buffer we use for data in and out on MMC */
776 host->data_mem = platform_get_resource(pdev, IORESOURCE_MEM, 1);
777 if (!host->data_mem) {
778 dev_err(&pdev->dev,
779 "failed to get io memory region resource.\n");
780 ret = -ENOENT;
781 goto probe_iounmap_mmio;
782 }
783
784 host->data_mem = request_mem_region(host->data_mem->start,
785 resource_size(host->data_mem),
786 pdev->name);
787
788 if (!host->data_mem) {
789 dev_err(&pdev->dev, "failed to request io memory region.\n");
790 ret = -ENOENT;
791 goto probe_iounmap_mmio;
792 }
793 host->data_base = ioremap(host->data_mem->start,
794 resource_size(host->data_mem));
795
796 if (host->data_base == 0) {
797 dev_err(&pdev->dev, "failed to ioremap() io memory region.\n");
798 ret = -EINVAL;
799 goto probe_free_mem_region_data;
800 }
801
802 ret = request_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), glamo_mci_irq, IRQF_SHARED,
803 pdev->name, host);
804 if (ret) {
805 dev_err(&pdev->dev, "failed to register irq.\n");
806 goto probe_iounmap_data;
807 }
808
809
810 host->vdd_current = -1;
811 host->clk_rate = 50000000; /* really it's 49152000 */
812 host->clk_div = 16;
813
814 /* explain our host controller capabilities */
815 mmc->ops = &glamo_mci_ops;
816 mmc->ocr_avail = mmc_regulator_get_ocrmask(host->regulator);
817 mmc->caps = MMC_CAP_4_BIT_DATA |
818 MMC_CAP_MMC_HIGHSPEED |
819 MMC_CAP_SD_HIGHSPEED;
820 mmc->f_min = host->clk_rate / 256;
821 mmc->f_max = sd_max_clk;
822
823 mmc->max_blk_count = (1 << 16) - 1; /* GLAMO_REG_MMC_RB_BLKCNT */
824 mmc->max_blk_size = (1 << 12) - 1; /* GLAMO_REG_MMC_RB_BLKLEN */
825 mmc->max_req_size = resource_size(host->data_mem);
826 mmc->max_seg_size = mmc->max_req_size;
827 mmc->max_phys_segs = 128;
828 mmc->max_hw_segs = 128;
829
830 platform_set_drvdata(pdev, mmc);
831
832 glamo_engine_enable(host->pdata->pglamo, GLAMO_ENGINE_MMC);
833 glamo_engine_reset(host->pdata->pglamo, GLAMO_ENGINE_MMC);
834
835 if ((ret = mmc_add_host(mmc))) {
836 dev_err(&pdev->dev, "failed to add mmc host.\n");
837 goto probe_freeirq;
838 }
839
840 writew((u16)(host->data_mem->start),
841 host->mmio_base + GLAMO_REG_MMC_WDATADS1);
842 writew((u16)((host->data_mem->start) >> 16),
843 host->mmio_base + GLAMO_REG_MMC_WDATADS2);
844
845 writew((u16)host->data_mem->start, host->mmio_base +
846 GLAMO_REG_MMC_RDATADS1);
847 writew((u16)(host->data_mem->start >> 16), host->mmio_base +
848 GLAMO_REG_MMC_RDATADS2);
849
850 dev_info(&pdev->dev,"initialisation done.\n");
851 return 0;
852
853 probe_freeirq:
854 free_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), host);
855 probe_iounmap_data:
856 iounmap(host->data_base);
857 probe_free_mem_region_data:
858 release_mem_region(host->data_mem->start, resource_size(host->data_mem));
859 probe_iounmap_mmio:
860 iounmap(host->mmio_base);
861 probe_free_mem_region_mmio:
862 release_mem_region(host->mmio_mem->start, resource_size(host->mmio_mem));
863 probe_regulator_put:
864 regulator_put(host->regulator);
865 probe_free_host:
866 mmc_free_host(mmc);
867 probe_out:
868 return ret;
869 }
870
871 static int glamo_mci_remove(struct platform_device *pdev)
872 {
873 struct mmc_host *mmc = platform_get_drvdata(pdev);
874 struct glamo_mci_host *host = mmc_priv(mmc);
875
876 free_irq(IRQ_GLAMO(GLAMO_IRQIDX_MMC), host);
877
878 mmc_remove_host(mmc);
879 iounmap(host->mmio_base);
880 iounmap(host->data_base);
881 release_mem_region(host->mmio_mem->start, resource_size(host->mmio_mem));
882 release_mem_region(host->data_mem->start, resource_size(host->data_mem));
883
884 regulator_put(host->regulator);
885
886 mmc_free_host(mmc);
887
888 glamo_engine_disable(host->pdata->pglamo, GLAMO_ENGINE_MMC);
889 return 0;
890 }
891
892
893 #ifdef CONFIG_PM
894
895 static int glamo_mci_suspend(struct platform_device *dev, pm_message_t state)
896 {
897 struct mmc_host *mmc = platform_get_drvdata(dev);
898 struct glamo_mci_host *host = mmc_priv(mmc);
899 int ret;
900
901 cancel_work_sync(&host->irq_work);
902
903 /*
904 * possible workaround for SD corruption during suspend - resume
905 * make sure the clock was running during suspend and consequently
906 * resume
907 */
908 glamo_mci_fix_card_div(host, host->clk_div);
909
910 /* we are going to do more commands to override this in
911 * mmc_suspend_host(), so we need to change sd_idleclk for the
912 * duration as well
913 */
914 suspend_sd_idleclk = sd_idleclk;
915 sd_idleclk = 1;
916
917 ret = mmc_suspend_host(mmc, state);
918
919 host->suspending++;
920
921 return ret;
922 }
923
924 int glamo_mci_resume(struct platform_device *dev)
925 {
926 struct mmc_host *mmc = platform_get_drvdata(dev);
927 struct glamo_mci_host *host = mmc_priv(mmc);
928 int ret;
929
930 sd_idleclk = 1;
931
932 glamo_engine_enable(host->pdata->pglamo, GLAMO_ENGINE_MMC);
933 glamo_engine_reset(host->pdata->pglamo, GLAMO_ENGINE_MMC);
934
935 host->suspending--;
936
937 ret = mmc_resume_host(mmc);
938
939 /* put sd_idleclk back to pre-suspend state */
940 sd_idleclk = suspend_sd_idleclk;
941
942 return ret;
943 }
944 EXPORT_SYMBOL_GPL(glamo_mci_resume);
945
946 #else /* CONFIG_PM */
947 #define glamo_mci_suspend NULL
948 #define glamo_mci_resume NULL
949 #endif /* CONFIG_PM */
950
951
952 static struct platform_driver glamo_mci_driver =
953 {
954 .driver.name = "glamo-mci",
955 .probe = glamo_mci_probe,
956 .remove = glamo_mci_remove,
957 .suspend = glamo_mci_suspend,
958 .resume = glamo_mci_resume,
959 };
960
961 static int __init glamo_mci_init(void)
962 {
963 platform_driver_register(&glamo_mci_driver);
964 return 0;
965 }
966
967 static void __exit glamo_mci_exit(void)
968 {
969 platform_driver_unregister(&glamo_mci_driver);
970 }
971
972 module_init(glamo_mci_init);
973 module_exit(glamo_mci_exit);
974
975 MODULE_DESCRIPTION("Glamo MMC/SD Card Interface driver");
976 MODULE_LICENSE("GPL");
977 MODULE_AUTHOR("Andy Green <andy@openmoko.com>");