1 /* 2 * Copyright 2011, Marvell Semiconductor Inc. 3 * Lei Wen <leiwen@marvell.com> 4 * 5 * SPDX-License-Identifier: GPL-2.0+ 6 * 7 * Back ported to the 8xx platform (from the 8260 platform) by 8 * Murray.Jensen@cmst.csiro.au, 27-Jan-01. 9 */ 10 11 #include <common.h> 12 #include <command.h> 13 #include <config.h> 14 #include <net.h> 15 #include <malloc.h> 16 #include <asm/byteorder.h> 17 #include <asm/errno.h> 18 #include <asm/io.h> 19 #include <asm/unaligned.h> 20 #include <linux/types.h> 21 #include <linux/usb/ch9.h> 22 #include <linux/usb/gadget.h> 23 #include <usb/ci_udc.h> 24 #include "../host/ehci.h" 25 #include "ci_udc.h" 26 27 /* 28 * Check if the system has too long cachelines. If the cachelines are 29 * longer then 128b, the driver will not be able flush/invalidate data 30 * cache over separate QH entries. We use 128b because one QH entry is 31 * 64b long and there are always two QH list entries for each endpoint. 32 */ 33 #if ARCH_DMA_MINALIGN > 128 34 #error This driver can not work on systems with caches longer than 128b 35 #endif 36 37 /* 38 * Every QTD must be individually aligned, since we can program any 39 * QTD's address into HW. Cache flushing requires ARCH_DMA_MINALIGN, 40 * and the USB HW requires 32-byte alignment. Align to both: 41 */ 42 #define ILIST_ALIGN roundup(ARCH_DMA_MINALIGN, 32) 43 /* Each QTD is this size */ 44 #define ILIST_ENT_RAW_SZ sizeof(struct ept_queue_item) 45 /* 46 * Align the size of the QTD too, so we can add this value to each 47 * QTD's address to get another aligned address. 48 */ 49 #define ILIST_ENT_SZ roundup(ILIST_ENT_RAW_SZ, ILIST_ALIGN) 50 /* For each endpoint, we need 2 QTDs, one for each of IN and OUT */ 51 #define ILIST_SZ (NUM_ENDPOINTS * 2 * ILIST_ENT_SZ) 52 53 #define EP_MAX_LENGTH_TRANSFER 0x4000 54 55 #ifndef DEBUG 56 #define DBG(x...) do {} while (0) 57 #else 58 #define DBG(x...) printf(x) 59 static const char *reqname(unsigned r) 60 { 61 switch (r) { 62 case USB_REQ_GET_STATUS: return "GET_STATUS"; 63 case USB_REQ_CLEAR_FEATURE: return "CLEAR_FEATURE"; 64 case USB_REQ_SET_FEATURE: return "SET_FEATURE"; 65 case USB_REQ_SET_ADDRESS: return "SET_ADDRESS"; 66 case USB_REQ_GET_DESCRIPTOR: return "GET_DESCRIPTOR"; 67 case USB_REQ_SET_DESCRIPTOR: return "SET_DESCRIPTOR"; 68 case USB_REQ_GET_CONFIGURATION: return "GET_CONFIGURATION"; 69 case USB_REQ_SET_CONFIGURATION: return "SET_CONFIGURATION"; 70 case USB_REQ_GET_INTERFACE: return "GET_INTERFACE"; 71 case USB_REQ_SET_INTERFACE: return "SET_INTERFACE"; 72 default: return "*UNKNOWN*"; 73 } 74 } 75 #endif 76 77 static struct usb_endpoint_descriptor ep0_desc = { 78 .bLength = sizeof(struct usb_endpoint_descriptor), 79 .bDescriptorType = USB_DT_ENDPOINT, 80 .bEndpointAddress = USB_DIR_IN, 81 .bmAttributes = USB_ENDPOINT_XFER_CONTROL, 82 }; 83 84 static int ci_pullup(struct usb_gadget *gadget, int is_on); 85 static int ci_ep_enable(struct usb_ep *ep, 86 const struct usb_endpoint_descriptor *desc); 87 static int ci_ep_disable(struct usb_ep *ep); 88 static int ci_ep_queue(struct usb_ep *ep, 89 struct usb_request *req, gfp_t gfp_flags); 90 static int ci_ep_dequeue(struct usb_ep *ep, struct usb_request *req); 91 static struct usb_request * 92 ci_ep_alloc_request(struct usb_ep *ep, unsigned int gfp_flags); 93 static void ci_ep_free_request(struct usb_ep *ep, struct usb_request *_req); 94 95 static struct usb_gadget_ops ci_udc_ops = { 96 .pullup = ci_pullup, 97 }; 98 99 static struct usb_ep_ops ci_ep_ops = { 100 .enable = ci_ep_enable, 101 .disable = ci_ep_disable, 102 .queue = ci_ep_queue, 103 .dequeue = ci_ep_dequeue, 104 .alloc_request = ci_ep_alloc_request, 105 .free_request = ci_ep_free_request, 106 }; 107 108 /* Init values for USB endpoints. */ 109 static const struct usb_ep ci_ep_init[5] = { 110 [0] = { /* EP 0 */ 111 .maxpacket = 64, 112 .name = "ep0", 113 .ops = &ci_ep_ops, 114 }, 115 [1] = { 116 .maxpacket = 512, 117 .name = "ep1in-bulk", 118 .ops = &ci_ep_ops, 119 }, 120 [2] = { 121 .maxpacket = 512, 122 .name = "ep2out-bulk", 123 .ops = &ci_ep_ops, 124 }, 125 [3] = { 126 .maxpacket = 512, 127 .name = "ep3in-int", 128 .ops = &ci_ep_ops, 129 }, 130 [4] = { 131 .maxpacket = 512, 132 .name = "ep-", 133 .ops = &ci_ep_ops, 134 }, 135 }; 136 137 static struct ci_drv controller = { 138 .gadget = { 139 .name = "ci_udc", 140 .ops = &ci_udc_ops, 141 .is_dualspeed = 1, 142 }, 143 }; 144 145 /** 146 * ci_get_qh() - return queue head for endpoint 147 * @ep_num: Endpoint number 148 * @dir_in: Direction of the endpoint (IN = 1, OUT = 0) 149 * 150 * This function returns the QH associated with particular endpoint 151 * and it's direction. 152 */ 153 static struct ept_queue_head *ci_get_qh(int ep_num, int dir_in) 154 { 155 return &controller.epts[(ep_num * 2) + dir_in]; 156 } 157 158 /** 159 * ci_get_qtd() - return queue item for endpoint 160 * @ep_num: Endpoint number 161 * @dir_in: Direction of the endpoint (IN = 1, OUT = 0) 162 * 163 * This function returns the QH associated with particular endpoint 164 * and it's direction. 165 */ 166 static struct ept_queue_item *ci_get_qtd(int ep_num, int dir_in) 167 { 168 int index = (ep_num * 2) + dir_in; 169 uint8_t *imem = controller.items_mem + (index * ILIST_ENT_SZ); 170 return (struct ept_queue_item *)imem; 171 } 172 173 /** 174 * ci_flush_qh - flush cache over queue head 175 * @ep_num: Endpoint number 176 * 177 * This function flushes cache over QH for particular endpoint. 178 */ 179 static void ci_flush_qh(int ep_num) 180 { 181 struct ept_queue_head *head = ci_get_qh(ep_num, 0); 182 const unsigned long start = (unsigned long)head; 183 const unsigned long end = start + 2 * sizeof(*head); 184 185 flush_dcache_range(start, end); 186 } 187 188 /** 189 * ci_invalidate_qh - invalidate cache over queue head 190 * @ep_num: Endpoint number 191 * 192 * This function invalidates cache over QH for particular endpoint. 193 */ 194 static void ci_invalidate_qh(int ep_num) 195 { 196 struct ept_queue_head *head = ci_get_qh(ep_num, 0); 197 unsigned long start = (unsigned long)head; 198 unsigned long end = start + 2 * sizeof(*head); 199 200 invalidate_dcache_range(start, end); 201 } 202 203 /** 204 * ci_flush_qtd - flush cache over queue item 205 * @ep_num: Endpoint number 206 * 207 * This function flushes cache over qTD pair for particular endpoint. 208 */ 209 static void ci_flush_qtd(int ep_num) 210 { 211 struct ept_queue_item *item = ci_get_qtd(ep_num, 0); 212 const unsigned long start = (unsigned long)item; 213 const unsigned long end = start + 2 * ILIST_ENT_SZ; 214 215 flush_dcache_range(start, end); 216 } 217 218 /** 219 * ci_flush_td - flush cache over queue item 220 * @td: td pointer 221 * 222 * This function flushes cache for particular transfer descriptor. 223 */ 224 static void ci_flush_td(struct ept_queue_item *td) 225 { 226 const unsigned long start = (unsigned long)td; 227 const unsigned long end = (unsigned long)td + ILIST_ENT_SZ; 228 flush_dcache_range(start, end); 229 } 230 231 /** 232 * ci_invalidate_qtd - invalidate cache over queue item 233 * @ep_num: Endpoint number 234 * 235 * This function invalidates cache over qTD pair for particular endpoint. 236 */ 237 static void ci_invalidate_qtd(int ep_num) 238 { 239 struct ept_queue_item *item = ci_get_qtd(ep_num, 0); 240 const unsigned long start = (unsigned long)item; 241 const unsigned long end = start + 2 * ILIST_ENT_SZ; 242 243 invalidate_dcache_range(start, end); 244 } 245 246 /** 247 * ci_invalidate_td - invalidate cache over queue item 248 * @td: td pointer 249 * 250 * This function invalidates cache for particular transfer descriptor. 251 */ 252 static void ci_invalidate_td(struct ept_queue_item *td) 253 { 254 const unsigned long start = (unsigned long)td; 255 const unsigned long end = start + ILIST_ENT_SZ; 256 invalidate_dcache_range(start, end); 257 } 258 259 static struct usb_request * 260 ci_ep_alloc_request(struct usb_ep *ep, unsigned int gfp_flags) 261 { 262 struct ci_ep *ci_ep = container_of(ep, struct ci_ep, ep); 263 int num = -1; 264 struct ci_req *ci_req; 265 266 if (ci_ep->desc) 267 num = ci_ep->desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK; 268 269 if (num == 0 && controller.ep0_req) 270 return &controller.ep0_req->req; 271 272 ci_req = calloc(1, sizeof(*ci_req)); 273 if (!ci_req) 274 return NULL; 275 276 INIT_LIST_HEAD(&ci_req->queue); 277 278 if (num == 0) 279 controller.ep0_req = ci_req; 280 281 return &ci_req->req; 282 } 283 284 static void ci_ep_free_request(struct usb_ep *ep, struct usb_request *req) 285 { 286 struct ci_ep *ci_ep = container_of(ep, struct ci_ep, ep); 287 struct ci_req *ci_req = container_of(req, struct ci_req, req); 288 int num = -1; 289 290 if (ci_ep->desc) 291 num = ci_ep->desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK; 292 293 if (num == 0) { 294 if (!controller.ep0_req) 295 return; 296 controller.ep0_req = 0; 297 } 298 299 if (ci_req->b_buf) 300 free(ci_req->b_buf); 301 free(ci_req); 302 } 303 304 static void ep_enable(int num, int in, int maxpacket) 305 { 306 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 307 unsigned n; 308 309 n = readl(&udc->epctrl[num]); 310 if (in) 311 n |= (CTRL_TXE | CTRL_TXR | CTRL_TXT_BULK); 312 else 313 n |= (CTRL_RXE | CTRL_RXR | CTRL_RXT_BULK); 314 315 if (num != 0) { 316 struct ept_queue_head *head = ci_get_qh(num, in); 317 318 head->config = CONFIG_MAX_PKT(maxpacket) | CONFIG_ZLT; 319 ci_flush_qh(num); 320 } 321 writel(n, &udc->epctrl[num]); 322 } 323 324 static int ci_ep_enable(struct usb_ep *ep, 325 const struct usb_endpoint_descriptor *desc) 326 { 327 struct ci_ep *ci_ep = container_of(ep, struct ci_ep, ep); 328 int num, in; 329 num = desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK; 330 in = (desc->bEndpointAddress & USB_DIR_IN) != 0; 331 ci_ep->desc = desc; 332 333 if (num) { 334 int max = get_unaligned_le16(&desc->wMaxPacketSize); 335 336 if ((max > 64) && (controller.gadget.speed == USB_SPEED_FULL)) 337 max = 64; 338 if (ep->maxpacket != max) { 339 DBG("%s: from %d to %d\n", __func__, 340 ep->maxpacket, max); 341 ep->maxpacket = max; 342 } 343 } 344 ep_enable(num, in, ep->maxpacket); 345 DBG("%s: num=%d maxpacket=%d\n", __func__, num, ep->maxpacket); 346 return 0; 347 } 348 349 static int ci_ep_disable(struct usb_ep *ep) 350 { 351 struct ci_ep *ci_ep = container_of(ep, struct ci_ep, ep); 352 353 ci_ep->desc = NULL; 354 return 0; 355 } 356 357 static int ci_bounce(struct ci_req *ci_req, int in) 358 { 359 struct usb_request *req = &ci_req->req; 360 unsigned long addr = (unsigned long)req->buf; 361 unsigned long hwaddr; 362 uint32_t aligned_used_len; 363 364 /* Input buffer address is not aligned. */ 365 if (addr & (ARCH_DMA_MINALIGN - 1)) 366 goto align; 367 368 /* Input buffer length is not aligned. */ 369 if (req->length & (ARCH_DMA_MINALIGN - 1)) 370 goto align; 371 372 /* The buffer is well aligned, only flush cache. */ 373 ci_req->hw_len = req->length; 374 ci_req->hw_buf = req->buf; 375 goto flush; 376 377 align: 378 if (ci_req->b_buf && req->length > ci_req->b_len) { 379 free(ci_req->b_buf); 380 ci_req->b_buf = 0; 381 } 382 if (!ci_req->b_buf) { 383 ci_req->b_len = roundup(req->length, ARCH_DMA_MINALIGN); 384 ci_req->b_buf = memalign(ARCH_DMA_MINALIGN, ci_req->b_len); 385 if (!ci_req->b_buf) 386 return -ENOMEM; 387 } 388 ci_req->hw_len = ci_req->b_len; 389 ci_req->hw_buf = ci_req->b_buf; 390 391 if (in) 392 memcpy(ci_req->hw_buf, req->buf, req->length); 393 394 flush: 395 hwaddr = (unsigned long)ci_req->hw_buf; 396 aligned_used_len = roundup(req->length, ARCH_DMA_MINALIGN); 397 flush_dcache_range(hwaddr, hwaddr + aligned_used_len); 398 399 return 0; 400 } 401 402 static void ci_debounce(struct ci_req *ci_req, int in) 403 { 404 struct usb_request *req = &ci_req->req; 405 unsigned long addr = (unsigned long)req->buf; 406 unsigned long hwaddr = (unsigned long)ci_req->hw_buf; 407 uint32_t aligned_used_len; 408 409 if (in) 410 return; 411 412 aligned_used_len = roundup(req->actual, ARCH_DMA_MINALIGN); 413 invalidate_dcache_range(hwaddr, hwaddr + aligned_used_len); 414 415 if (addr == hwaddr) 416 return; /* not a bounce */ 417 418 memcpy(req->buf, ci_req->hw_buf, req->actual); 419 } 420 421 static void ci_ep_submit_next_request(struct ci_ep *ci_ep) 422 { 423 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 424 struct ept_queue_item *item; 425 struct ept_queue_head *head; 426 int bit, num, len, in; 427 struct ci_req *ci_req; 428 u8 *buf; 429 uint32_t length, actlen; 430 struct ept_queue_item *dtd, *qtd; 431 432 ci_ep->req_primed = true; 433 434 num = ci_ep->desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK; 435 in = (ci_ep->desc->bEndpointAddress & USB_DIR_IN) != 0; 436 item = ci_get_qtd(num, in); 437 head = ci_get_qh(num, in); 438 439 ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue); 440 len = ci_req->req.length; 441 442 head->next = (unsigned long)item; 443 head->info = 0; 444 445 ci_req->dtd_count = 0; 446 buf = ci_req->hw_buf; 447 actlen = 0; 448 dtd = item; 449 450 do { 451 length = min(ci_req->req.length - actlen, 452 (unsigned)EP_MAX_LENGTH_TRANSFER); 453 454 dtd->info = INFO_BYTES(length) | INFO_ACTIVE; 455 dtd->page0 = (unsigned long)buf; 456 dtd->page1 = ((unsigned long)buf & 0xfffff000) + 0x1000; 457 dtd->page2 = ((unsigned long)buf & 0xfffff000) + 0x2000; 458 dtd->page3 = ((unsigned long)buf & 0xfffff000) + 0x3000; 459 dtd->page4 = ((unsigned long)buf & 0xfffff000) + 0x4000; 460 461 len -= length; 462 actlen += length; 463 buf += length; 464 465 if (len) { 466 qtd = (struct ept_queue_item *) 467 memalign(ILIST_ALIGN, ILIST_ENT_SZ); 468 dtd->next = (unsigned long)qtd; 469 dtd = qtd; 470 memset(dtd, 0, ILIST_ENT_SZ); 471 } 472 473 ci_req->dtd_count++; 474 } while (len); 475 476 item = dtd; 477 /* 478 * When sending the data for an IN transaction, the attached host 479 * knows that all data for the IN is sent when one of the following 480 * occurs: 481 * a) A zero-length packet is transmitted. 482 * b) A packet with length that isn't an exact multiple of the ep's 483 * maxpacket is transmitted. 484 * c) Enough data is sent to exactly fill the host's maximum expected 485 * IN transaction size. 486 * 487 * One of these conditions MUST apply at the end of an IN transaction, 488 * or the transaction will not be considered complete by the host. If 489 * none of (a)..(c) already applies, then we must force (a) to apply 490 * by explicitly sending an extra zero-length packet. 491 */ 492 /* IN !a !b !c */ 493 if (in && len && !(len % ci_ep->ep.maxpacket) && ci_req->req.zero) { 494 /* 495 * Each endpoint has 2 items allocated, even though typically 496 * only 1 is used at a time since either an IN or an OUT but 497 * not both is queued. For an IN transaction, item currently 498 * points at the second of these items, so we know that we 499 * can use the other to transmit the extra zero-length packet. 500 */ 501 struct ept_queue_item *other_item = ci_get_qtd(num, 0); 502 item->next = (unsigned long)other_item; 503 item = other_item; 504 item->info = INFO_ACTIVE; 505 } 506 507 item->next = TERMINATE; 508 item->info |= INFO_IOC; 509 510 ci_flush_qtd(num); 511 512 item = (struct ept_queue_item *)(unsigned long)head->next; 513 while (item->next != TERMINATE) { 514 ci_flush_td((struct ept_queue_item *)(unsigned long)item->next); 515 item = (struct ept_queue_item *)(unsigned long)item->next; 516 } 517 518 DBG("ept%d %s queue len %x, req %p, buffer %p\n", 519 num, in ? "in" : "out", len, ci_req, ci_req->hw_buf); 520 ci_flush_qh(num); 521 522 if (in) 523 bit = EPT_TX(num); 524 else 525 bit = EPT_RX(num); 526 527 writel(bit, &udc->epprime); 528 } 529 530 static int ci_ep_dequeue(struct usb_ep *_ep, struct usb_request *_req) 531 { 532 struct ci_ep *ci_ep = container_of(_ep, struct ci_ep, ep); 533 struct ci_req *ci_req; 534 535 list_for_each_entry(ci_req, &ci_ep->queue, queue) { 536 if (&ci_req->req == _req) 537 break; 538 } 539 540 if (&ci_req->req != _req) 541 return -EINVAL; 542 543 list_del_init(&ci_req->queue); 544 545 if (ci_req->req.status == -EINPROGRESS) { 546 ci_req->req.status = -ECONNRESET; 547 if (ci_req->req.complete) 548 ci_req->req.complete(_ep, _req); 549 } 550 551 return 0; 552 } 553 554 static int ci_ep_queue(struct usb_ep *ep, 555 struct usb_request *req, gfp_t gfp_flags) 556 { 557 struct ci_ep *ci_ep = container_of(ep, struct ci_ep, ep); 558 struct ci_req *ci_req = container_of(req, struct ci_req, req); 559 int in, ret; 560 int __maybe_unused num; 561 562 num = ci_ep->desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK; 563 in = (ci_ep->desc->bEndpointAddress & USB_DIR_IN) != 0; 564 565 if (!num && ci_ep->req_primed) { 566 /* 567 * The flipping of ep0 between IN and OUT relies on 568 * ci_ep_queue consuming the current IN/OUT setting 569 * immediately. If this is deferred to a later point when the 570 * req is pulled out of ci_req->queue, then the IN/OUT setting 571 * may have been changed since the req was queued, and state 572 * will get out of sync. This condition doesn't occur today, 573 * but could if bugs were introduced later, and this error 574 * check will save a lot of debugging time. 575 */ 576 printf("%s: ep0 transaction already in progress\n", __func__); 577 return -EPROTO; 578 } 579 580 ret = ci_bounce(ci_req, in); 581 if (ret) 582 return ret; 583 584 DBG("ept%d %s pre-queue req %p, buffer %p\n", 585 num, in ? "in" : "out", ci_req, ci_req->hw_buf); 586 list_add_tail(&ci_req->queue, &ci_ep->queue); 587 588 if (!ci_ep->req_primed) 589 ci_ep_submit_next_request(ci_ep); 590 591 return 0; 592 } 593 594 static void flip_ep0_direction(void) 595 { 596 if (ep0_desc.bEndpointAddress == USB_DIR_IN) { 597 DBG("%s: Flipping ep0 to OUT\n", __func__); 598 ep0_desc.bEndpointAddress = 0; 599 } else { 600 DBG("%s: Flipping ep0 to IN\n", __func__); 601 ep0_desc.bEndpointAddress = USB_DIR_IN; 602 } 603 } 604 605 static void handle_ep_complete(struct ci_ep *ci_ep) 606 { 607 struct ept_queue_item *item, *next_td; 608 int num, in, len, j; 609 struct ci_req *ci_req; 610 611 num = ci_ep->desc->bEndpointAddress & USB_ENDPOINT_NUMBER_MASK; 612 in = (ci_ep->desc->bEndpointAddress & USB_DIR_IN) != 0; 613 item = ci_get_qtd(num, in); 614 ci_invalidate_qtd(num); 615 ci_req = list_first_entry(&ci_ep->queue, struct ci_req, queue); 616 617 next_td = item; 618 len = 0; 619 for (j = 0; j < ci_req->dtd_count; j++) { 620 ci_invalidate_td(next_td); 621 item = next_td; 622 len += (item->info >> 16) & 0x7fff; 623 if (item->info & 0xff) 624 printf("EP%d/%s FAIL info=%x pg0=%x\n", 625 num, in ? "in" : "out", item->info, item->page0); 626 if (j != ci_req->dtd_count - 1) 627 next_td = (struct ept_queue_item *)(unsigned long) 628 item->next; 629 if (j != 0) 630 free(item); 631 } 632 633 list_del_init(&ci_req->queue); 634 ci_ep->req_primed = false; 635 636 if (!list_empty(&ci_ep->queue)) 637 ci_ep_submit_next_request(ci_ep); 638 639 ci_req->req.actual = ci_req->req.length - len; 640 ci_debounce(ci_req, in); 641 642 DBG("ept%d %s req %p, complete %x\n", 643 num, in ? "in" : "out", ci_req, len); 644 if (num != 0 || controller.ep0_data_phase) 645 ci_req->req.complete(&ci_ep->ep, &ci_req->req); 646 if (num == 0 && controller.ep0_data_phase) { 647 /* 648 * Data Stage is complete, so flip ep0 dir for Status Stage, 649 * which always transfers a packet in the opposite direction. 650 */ 651 DBG("%s: flip ep0 dir for Status Stage\n", __func__); 652 flip_ep0_direction(); 653 controller.ep0_data_phase = false; 654 ci_req->req.length = 0; 655 usb_ep_queue(&ci_ep->ep, &ci_req->req, 0); 656 } 657 } 658 659 #define SETUP(type, request) (((type) << 8) | (request)) 660 661 static void handle_setup(void) 662 { 663 struct ci_ep *ci_ep = &controller.ep[0]; 664 struct ci_req *ci_req; 665 struct usb_request *req; 666 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 667 struct ept_queue_head *head; 668 struct usb_ctrlrequest r; 669 int status = 0; 670 int num, in, _num, _in, i; 671 char *buf; 672 673 ci_req = controller.ep0_req; 674 req = &ci_req->req; 675 head = ci_get_qh(0, 0); /* EP0 OUT */ 676 677 ci_invalidate_qh(0); 678 memcpy(&r, head->setup_data, sizeof(struct usb_ctrlrequest)); 679 #ifdef CONFIG_CI_UDC_HAS_HOSTPC 680 writel(EPT_RX(0), &udc->epsetupstat); 681 #else 682 writel(EPT_RX(0), &udc->epstat); 683 #endif 684 DBG("handle setup %s, %x, %x index %x value %x length %x\n", 685 reqname(r.bRequest), r.bRequestType, r.bRequest, r.wIndex, 686 r.wValue, r.wLength); 687 688 /* Set EP0 dir for Data Stage based on Setup Stage data */ 689 if (r.bRequestType & USB_DIR_IN) { 690 DBG("%s: Set ep0 to IN for Data Stage\n", __func__); 691 ep0_desc.bEndpointAddress = USB_DIR_IN; 692 } else { 693 DBG("%s: Set ep0 to OUT for Data Stage\n", __func__); 694 ep0_desc.bEndpointAddress = 0; 695 } 696 if (r.wLength) { 697 controller.ep0_data_phase = true; 698 } else { 699 /* 0 length -> no Data Stage. Flip dir for Status Stage */ 700 DBG("%s: 0 length: flip ep0 dir for Status Stage\n", __func__); 701 flip_ep0_direction(); 702 controller.ep0_data_phase = false; 703 } 704 705 list_del_init(&ci_req->queue); 706 ci_ep->req_primed = false; 707 708 switch (SETUP(r.bRequestType, r.bRequest)) { 709 case SETUP(USB_RECIP_ENDPOINT, USB_REQ_CLEAR_FEATURE): 710 _num = r.wIndex & 15; 711 _in = !!(r.wIndex & 0x80); 712 713 if ((r.wValue == 0) && (r.wLength == 0)) { 714 req->length = 0; 715 for (i = 0; i < NUM_ENDPOINTS; i++) { 716 struct ci_ep *ep = &controller.ep[i]; 717 718 if (!ep->desc) 719 continue; 720 num = ep->desc->bEndpointAddress 721 & USB_ENDPOINT_NUMBER_MASK; 722 in = (ep->desc->bEndpointAddress 723 & USB_DIR_IN) != 0; 724 if ((num == _num) && (in == _in)) { 725 ep_enable(num, in, ep->ep.maxpacket); 726 usb_ep_queue(controller.gadget.ep0, 727 req, 0); 728 break; 729 } 730 } 731 } 732 return; 733 734 case SETUP(USB_RECIP_DEVICE, USB_REQ_SET_ADDRESS): 735 /* 736 * write address delayed (will take effect 737 * after the next IN txn) 738 */ 739 writel((r.wValue << 25) | (1 << 24), &udc->devaddr); 740 req->length = 0; 741 usb_ep_queue(controller.gadget.ep0, req, 0); 742 return; 743 744 case SETUP(USB_DIR_IN | USB_RECIP_DEVICE, USB_REQ_GET_STATUS): 745 req->length = 2; 746 buf = (char *)req->buf; 747 buf[0] = 1 << USB_DEVICE_SELF_POWERED; 748 buf[1] = 0; 749 usb_ep_queue(controller.gadget.ep0, req, 0); 750 return; 751 } 752 /* pass request up to the gadget driver */ 753 if (controller.driver) 754 status = controller.driver->setup(&controller.gadget, &r); 755 else 756 status = -ENODEV; 757 758 if (!status) 759 return; 760 DBG("STALL reqname %s type %x value %x, index %x\n", 761 reqname(r.bRequest), r.bRequestType, r.wValue, r.wIndex); 762 writel((1<<16) | (1 << 0), &udc->epctrl[0]); 763 } 764 765 static void stop_activity(void) 766 { 767 int i, num, in; 768 struct ept_queue_head *head; 769 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 770 writel(readl(&udc->epcomp), &udc->epcomp); 771 #ifdef CONFIG_CI_UDC_HAS_HOSTPC 772 writel(readl(&udc->epsetupstat), &udc->epsetupstat); 773 #endif 774 writel(readl(&udc->epstat), &udc->epstat); 775 writel(0xffffffff, &udc->epflush); 776 777 /* error out any pending reqs */ 778 for (i = 0; i < NUM_ENDPOINTS; i++) { 779 if (i != 0) 780 writel(0, &udc->epctrl[i]); 781 if (controller.ep[i].desc) { 782 num = controller.ep[i].desc->bEndpointAddress 783 & USB_ENDPOINT_NUMBER_MASK; 784 in = (controller.ep[i].desc->bEndpointAddress 785 & USB_DIR_IN) != 0; 786 head = ci_get_qh(num, in); 787 head->info = INFO_ACTIVE; 788 ci_flush_qh(num); 789 } 790 } 791 } 792 793 void udc_irq(void) 794 { 795 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 796 unsigned n = readl(&udc->usbsts); 797 writel(n, &udc->usbsts); 798 int bit, i, num, in; 799 800 n &= (STS_SLI | STS_URI | STS_PCI | STS_UI | STS_UEI); 801 if (n == 0) 802 return; 803 804 if (n & STS_URI) { 805 DBG("-- reset --\n"); 806 stop_activity(); 807 } 808 if (n & STS_SLI) 809 DBG("-- suspend --\n"); 810 811 if (n & STS_PCI) { 812 int max = 64; 813 int speed = USB_SPEED_FULL; 814 815 #ifdef CONFIG_CI_UDC_HAS_HOSTPC 816 bit = (readl(&udc->hostpc1_devlc) >> 25) & 3; 817 #else 818 bit = (readl(&udc->portsc) >> 26) & 3; 819 #endif 820 DBG("-- portchange %x %s\n", bit, (bit == 2) ? "High" : "Full"); 821 if (bit == 2) { 822 speed = USB_SPEED_HIGH; 823 max = 512; 824 } 825 controller.gadget.speed = speed; 826 for (i = 1; i < NUM_ENDPOINTS; i++) { 827 if (controller.ep[i].ep.maxpacket > max) 828 controller.ep[i].ep.maxpacket = max; 829 } 830 } 831 832 if (n & STS_UEI) 833 printf("<UEI %x>\n", readl(&udc->epcomp)); 834 835 if ((n & STS_UI) || (n & STS_UEI)) { 836 #ifdef CONFIG_CI_UDC_HAS_HOSTPC 837 n = readl(&udc->epsetupstat); 838 #else 839 n = readl(&udc->epstat); 840 #endif 841 if (n & EPT_RX(0)) 842 handle_setup(); 843 844 n = readl(&udc->epcomp); 845 if (n != 0) 846 writel(n, &udc->epcomp); 847 848 for (i = 0; i < NUM_ENDPOINTS && n; i++) { 849 if (controller.ep[i].desc) { 850 num = controller.ep[i].desc->bEndpointAddress 851 & USB_ENDPOINT_NUMBER_MASK; 852 in = (controller.ep[i].desc->bEndpointAddress 853 & USB_DIR_IN) != 0; 854 bit = (in) ? EPT_TX(num) : EPT_RX(num); 855 if (n & bit) 856 handle_ep_complete(&controller.ep[i]); 857 } 858 } 859 } 860 } 861 862 int usb_gadget_handle_interrupts(int index) 863 { 864 u32 value; 865 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 866 867 value = readl(&udc->usbsts); 868 if (value) 869 udc_irq(); 870 871 return value; 872 } 873 874 void udc_disconnect(void) 875 { 876 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 877 /* disable pullup */ 878 stop_activity(); 879 writel(USBCMD_FS2, &udc->usbcmd); 880 udelay(800); 881 if (controller.driver) 882 controller.driver->disconnect(&controller.gadget); 883 } 884 885 static int ci_pullup(struct usb_gadget *gadget, int is_on) 886 { 887 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 888 if (is_on) { 889 /* RESET */ 890 writel(USBCMD_ITC(MICRO_8FRAME) | USBCMD_RST, &udc->usbcmd); 891 udelay(200); 892 893 writel((unsigned long)controller.epts, &udc->epinitaddr); 894 895 /* select DEVICE mode */ 896 writel(USBMODE_DEVICE, &udc->usbmode); 897 898 #if !defined(CONFIG_USB_GADGET_DUALSPEED) 899 /* Port force Full-Speed Connect */ 900 setbits_le32(&udc->portsc, PFSC); 901 #endif 902 903 writel(0xffffffff, &udc->epflush); 904 905 /* Turn on the USB connection by enabling the pullup resistor */ 906 writel(USBCMD_ITC(MICRO_8FRAME) | USBCMD_RUN, &udc->usbcmd); 907 } else { 908 udc_disconnect(); 909 } 910 911 return 0; 912 } 913 914 static int ci_udc_probe(void) 915 { 916 struct ept_queue_head *head; 917 int i; 918 919 const int num = 2 * NUM_ENDPOINTS; 920 921 const int eplist_min_align = 4096; 922 const int eplist_align = roundup(eplist_min_align, ARCH_DMA_MINALIGN); 923 const int eplist_raw_sz = num * sizeof(struct ept_queue_head); 924 const int eplist_sz = roundup(eplist_raw_sz, ARCH_DMA_MINALIGN); 925 926 /* The QH list must be aligned to 4096 bytes. */ 927 controller.epts = memalign(eplist_align, eplist_sz); 928 if (!controller.epts) 929 return -ENOMEM; 930 memset(controller.epts, 0, eplist_sz); 931 932 controller.items_mem = memalign(ILIST_ALIGN, ILIST_SZ); 933 if (!controller.items_mem) { 934 free(controller.epts); 935 return -ENOMEM; 936 } 937 memset(controller.items_mem, 0, ILIST_SZ); 938 939 for (i = 0; i < 2 * NUM_ENDPOINTS; i++) { 940 /* 941 * Configure QH for each endpoint. The structure of the QH list 942 * is such that each two subsequent fields, N and N+1 where N is 943 * even, in the QH list represent QH for one endpoint. The Nth 944 * entry represents OUT configuration and the N+1th entry does 945 * represent IN configuration of the endpoint. 946 */ 947 head = controller.epts + i; 948 if (i < 2) 949 head->config = CONFIG_MAX_PKT(EP0_MAX_PACKET_SIZE) 950 | CONFIG_ZLT | CONFIG_IOS; 951 else 952 head->config = CONFIG_MAX_PKT(EP_MAX_PACKET_SIZE) 953 | CONFIG_ZLT; 954 head->next = TERMINATE; 955 head->info = 0; 956 957 if (i & 1) { 958 ci_flush_qh(i / 2); 959 ci_flush_qtd(i / 2); 960 } 961 } 962 963 INIT_LIST_HEAD(&controller.gadget.ep_list); 964 965 /* Init EP 0 */ 966 memcpy(&controller.ep[0].ep, &ci_ep_init[0], sizeof(*ci_ep_init)); 967 controller.ep[0].desc = &ep0_desc; 968 INIT_LIST_HEAD(&controller.ep[0].queue); 969 controller.ep[0].req_primed = false; 970 controller.gadget.ep0 = &controller.ep[0].ep; 971 INIT_LIST_HEAD(&controller.gadget.ep0->ep_list); 972 973 /* Init EP 1..3 */ 974 for (i = 1; i < 4; i++) { 975 memcpy(&controller.ep[i].ep, &ci_ep_init[i], 976 sizeof(*ci_ep_init)); 977 INIT_LIST_HEAD(&controller.ep[i].queue); 978 controller.ep[i].req_primed = false; 979 list_add_tail(&controller.ep[i].ep.ep_list, 980 &controller.gadget.ep_list); 981 } 982 983 /* Init EP 4..n */ 984 for (i = 4; i < NUM_ENDPOINTS; i++) { 985 memcpy(&controller.ep[i].ep, &ci_ep_init[4], 986 sizeof(*ci_ep_init)); 987 INIT_LIST_HEAD(&controller.ep[i].queue); 988 controller.ep[i].req_primed = false; 989 list_add_tail(&controller.ep[i].ep.ep_list, 990 &controller.gadget.ep_list); 991 } 992 993 ci_ep_alloc_request(&controller.ep[0].ep, 0); 994 if (!controller.ep0_req) { 995 free(controller.items_mem); 996 free(controller.epts); 997 return -ENOMEM; 998 } 999 1000 return 0; 1001 } 1002 1003 int usb_gadget_register_driver(struct usb_gadget_driver *driver) 1004 { 1005 int ret; 1006 1007 if (!driver) 1008 return -EINVAL; 1009 if (!driver->bind || !driver->setup || !driver->disconnect) 1010 return -EINVAL; 1011 if (driver->speed != USB_SPEED_FULL && driver->speed != USB_SPEED_HIGH) 1012 return -EINVAL; 1013 1014 #ifdef CONFIG_DM_USB 1015 ret = usb_setup_ehci_gadget(&controller.ctrl); 1016 #else 1017 ret = usb_lowlevel_init(0, USB_INIT_DEVICE, (void **)&controller.ctrl); 1018 #endif 1019 if (ret) 1020 return ret; 1021 1022 ret = ci_udc_probe(); 1023 #if defined(CONFIG_USB_EHCI_MX6) || defined(CONFIG_USB_EHCI_MXS) 1024 /* 1025 * FIXME: usb_lowlevel_init()->ehci_hcd_init() should be doing all 1026 * HW-specific initialization, e.g. ULPI-vs-UTMI PHY selection 1027 */ 1028 if (!ret) { 1029 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 1030 1031 /* select ULPI phy */ 1032 writel(PTS(PTS_ENABLE) | PFSC, &udc->portsc); 1033 } 1034 #endif 1035 1036 ret = driver->bind(&controller.gadget); 1037 if (ret) { 1038 DBG("driver->bind() returned %d\n", ret); 1039 return ret; 1040 } 1041 controller.driver = driver; 1042 1043 return 0; 1044 } 1045 1046 int usb_gadget_unregister_driver(struct usb_gadget_driver *driver) 1047 { 1048 udc_disconnect(); 1049 1050 driver->unbind(&controller.gadget); 1051 controller.driver = NULL; 1052 1053 ci_ep_free_request(&controller.ep[0].ep, &controller.ep0_req->req); 1054 free(controller.items_mem); 1055 free(controller.epts); 1056 1057 return 0; 1058 } 1059 1060 bool dfu_usb_get_reset(void) 1061 { 1062 struct ci_udc *udc = (struct ci_udc *)controller.ctrl->hcor; 1063 1064 return !!(readl(&udc->usbsts) & STS_URI); 1065 } 1066