1 From 11527dad9862ba7e53654943fdacc3ffdad00ae2 Mon Sep 17 00:00:00 2001
2 From: Jonathan Bell <jonathan@raspberrypi.com>
3 Date: Mon, 13 Dec 2021 15:05:56 +0000
4 Subject: [PATCH] xhci: refactor out TRBS_PER_SEGMENT define in runtime
7 In anticipation of adjusting the number of utilised TRBs in a ring
8 segment, add trbs_per_seg to struct xhci_ring and use this instead
9 of a compile-time define.
11 Signed-off-by: Jonathan Bell <jonathan@raspberrypi.com>
13 drivers/usb/host/xhci-mem.c | 48 +++++++++++++++++++-----------------
14 drivers/usb/host/xhci-ring.c | 20 +++++++++------
15 drivers/usb/host/xhci.c | 6 ++---
16 drivers/usb/host/xhci.h | 1 +
17 4 files changed, 42 insertions(+), 33 deletions(-)
19 --- a/drivers/usb/host/xhci-mem.c
20 +++ b/drivers/usb/host/xhci-mem.c
21 @@ -98,6 +98,7 @@ static void xhci_free_segments_for_ring(
23 static void xhci_link_segments(struct xhci_segment *prev,
24 struct xhci_segment *next,
25 + unsigned int trbs_per_seg,
26 enum xhci_ring_type type, bool chain_links)
29 @@ -106,16 +107,16 @@ static void xhci_link_segments(struct xh
32 if (type != TYPE_EVENT) {
33 - prev->trbs[TRBS_PER_SEGMENT-1].link.segment_ptr =
34 + prev->trbs[trbs_per_seg - 1].link.segment_ptr =
35 cpu_to_le64(next->dma);
37 /* Set the last TRB in the segment to have a TRB type ID of Link TRB */
38 - val = le32_to_cpu(prev->trbs[TRBS_PER_SEGMENT-1].link.control);
39 + val = le32_to_cpu(prev->trbs[trbs_per_seg - 1].link.control);
40 val &= ~TRB_TYPE_BITMASK;
41 val |= TRB_TYPE(TRB_LINK);
44 - prev->trbs[TRBS_PER_SEGMENT-1].link.control = cpu_to_le32(val);
45 + prev->trbs[trbs_per_seg - 1].link.control = cpu_to_le32(val);
49 @@ -139,15 +140,17 @@ static void xhci_link_rings(struct xhci_
50 (xhci->quirks & XHCI_AMD_0x96_HOST)));
52 next = ring->enq_seg->next;
53 - xhci_link_segments(ring->enq_seg, first, ring->type, chain_links);
54 - xhci_link_segments(last, next, ring->type, chain_links);
55 + xhci_link_segments(ring->enq_seg, first, ring->trbs_per_seg,
56 + ring->type, chain_links);
57 + xhci_link_segments(last, next, ring->trbs_per_seg,
58 + ring->type, chain_links);
59 ring->num_segs += num_segs;
60 - ring->num_trbs_free += (TRBS_PER_SEGMENT - 1) * num_segs;
61 + ring->num_trbs_free += (ring->trbs_per_seg - 1) * num_segs;
63 if (ring->type != TYPE_EVENT && ring->enq_seg == ring->last_seg) {
64 - ring->last_seg->trbs[TRBS_PER_SEGMENT-1].link.control
65 + ring->last_seg->trbs[ring->trbs_per_seg - 1].link.control
66 &= ~cpu_to_le32(LINK_TOGGLE);
67 - last->trbs[TRBS_PER_SEGMENT-1].link.control
68 + last->trbs[ring->trbs_per_seg - 1].link.control
69 |= cpu_to_le32(LINK_TOGGLE);
70 ring->last_seg = last;
72 @@ -314,14 +317,15 @@ void xhci_initialize_ring_info(struct xh
73 * Each segment has a link TRB, and leave an extra TRB for SW
76 - ring->num_trbs_free = ring->num_segs * (TRBS_PER_SEGMENT - 1) - 1;
77 + ring->num_trbs_free = ring->num_segs * (ring->trbs_per_seg - 1) - 1;
80 /* Allocate segments and link them for a ring */
81 static int xhci_alloc_segments_for_ring(struct xhci_hcd *xhci,
82 struct xhci_segment **first, struct xhci_segment **last,
83 - unsigned int num_segs, unsigned int cycle_state,
84 - enum xhci_ring_type type, unsigned int max_packet, gfp_t flags)
85 + unsigned int num_segs, unsigned int trbs_per_seg,
86 + unsigned int cycle_state, enum xhci_ring_type type,
87 + unsigned int max_packet, gfp_t flags)
89 struct xhci_segment *prev;
91 @@ -350,12 +354,12 @@ static int xhci_alloc_segments_for_ring(
95 - xhci_link_segments(prev, next, type, chain_links);
96 + xhci_link_segments(prev, next, trbs_per_seg, type, chain_links);
101 - xhci_link_segments(prev, *first, type, chain_links);
102 + xhci_link_segments(prev, *first, trbs_per_seg, type, chain_links);
106 @@ -387,16 +391,17 @@ struct xhci_ring *xhci_ring_alloc(struct
110 + ring->trbs_per_seg = TRBS_PER_SEGMENT;
111 ret = xhci_alloc_segments_for_ring(xhci, &ring->first_seg,
112 - &ring->last_seg, num_segs, cycle_state, type,
113 - max_packet, flags);
114 + &ring->last_seg, num_segs, ring->trbs_per_seg,
115 + cycle_state, type, max_packet, flags);
119 /* Only event ring does not use link TRB */
120 if (type != TYPE_EVENT) {
121 /* See section 4.9.2.1 and 6.4.4.1 */
122 - ring->last_seg->trbs[TRBS_PER_SEGMENT - 1].link.control |=
123 + ring->last_seg->trbs[ring->trbs_per_seg - 1].link.control |=
124 cpu_to_le32(LINK_TOGGLE);
126 xhci_initialize_ring_info(ring, cycle_state);
127 @@ -429,15 +434,14 @@ int xhci_ring_expansion(struct xhci_hcd
128 unsigned int num_segs_needed;
131 - num_segs_needed = (num_trbs + (TRBS_PER_SEGMENT - 1) - 1) /
132 - (TRBS_PER_SEGMENT - 1);
134 + num_segs_needed = (num_trbs + (ring->trbs_per_seg - 1) - 1) /
135 + (ring->trbs_per_seg - 1);
136 /* Allocate number of segments we needed, or double the ring size */
137 num_segs = max(ring->num_segs, num_segs_needed);
139 ret = xhci_alloc_segments_for_ring(xhci, &first, &last,
140 - num_segs, ring->cycle_state, ring->type,
141 - ring->bounce_buf_len, flags);
142 + num_segs, ring->trbs_per_seg, ring->cycle_state,
143 + ring->type, ring->bounce_buf_len, flags);
147 @@ -1811,7 +1815,7 @@ int xhci_alloc_erst(struct xhci_hcd *xhc
148 for (val = 0; val < evt_ring->num_segs; val++) {
149 entry = &erst->entries[val];
150 entry->seg_addr = cpu_to_le64(seg->dma);
151 - entry->seg_size = cpu_to_le32(TRBS_PER_SEGMENT);
152 + entry->seg_size = cpu_to_le32(evt_ring->trbs_per_seg);
156 --- a/drivers/usb/host/xhci-ring.c
157 +++ b/drivers/usb/host/xhci-ring.c
158 @@ -90,15 +90,16 @@ static bool trb_is_link(union xhci_trb *
159 return TRB_TYPE_LINK_LE32(trb->link.control);
162 -static bool last_trb_on_seg(struct xhci_segment *seg, union xhci_trb *trb)
163 +static bool last_trb_on_seg(struct xhci_segment *seg,
164 + unsigned int trbs_per_seg, union xhci_trb *trb)
166 - return trb == &seg->trbs[TRBS_PER_SEGMENT - 1];
167 + return trb == &seg->trbs[trbs_per_seg - 1];
170 static bool last_trb_on_ring(struct xhci_ring *ring,
171 struct xhci_segment *seg, union xhci_trb *trb)
173 - return last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg);
174 + return last_trb_on_seg(seg, ring->trbs_per_seg, trb) && (seg->next == ring->first_seg);
177 static bool link_trb_toggles_cycle(union xhci_trb *trb)
178 @@ -161,7 +162,8 @@ void inc_deq(struct xhci_hcd *xhci, stru
180 /* event ring doesn't have link trbs, check for last trb */
181 if (ring->type == TYPE_EVENT) {
182 - if (!last_trb_on_seg(ring->deq_seg, ring->dequeue)) {
183 + if (!last_trb_on_seg(ring->deq_seg, ring->trbs_per_seg,
188 @@ -174,7 +176,8 @@ void inc_deq(struct xhci_hcd *xhci, stru
190 /* All other rings have link trbs */
191 if (!trb_is_link(ring->dequeue)) {
192 - if (last_trb_on_seg(ring->deq_seg, ring->dequeue)) {
193 + if (last_trb_on_seg(ring->deq_seg, ring->trbs_per_seg,
195 xhci_warn(xhci, "Missing link TRB at end of segment\n");
198 @@ -225,7 +228,7 @@ static void inc_enq(struct xhci_hcd *xhc
199 if (!trb_is_link(ring->enqueue))
200 ring->num_trbs_free--;
202 - if (last_trb_on_seg(ring->enq_seg, ring->enqueue)) {
203 + if (last_trb_on_seg(ring->enq_seg, ring->trbs_per_seg, ring->enqueue)) {
204 xhci_err(xhci, "Tried to move enqueue past ring segment\n");
207 @@ -3150,7 +3153,7 @@ irqreturn_t xhci_irq(struct usb_hcd *hcd
208 * that clears the EHB.
210 while (xhci_handle_event(xhci) > 0) {
211 - if (event_loop++ < TRBS_PER_SEGMENT / 2)
212 + if (event_loop++ < xhci->event_ring->trbs_per_seg / 2)
214 xhci_update_erst_dequeue(xhci, event_ring_deq);
215 event_ring_deq = xhci->event_ring->dequeue;
216 @@ -3292,7 +3295,8 @@ static int prepare_ring(struct xhci_hcd
220 - if (last_trb_on_seg(ep_ring->enq_seg, ep_ring->enqueue)) {
221 + if (last_trb_on_seg(ep_ring->enq_seg, ep_ring->trbs_per_seg,
222 + ep_ring->enqueue)) {
223 xhci_warn(xhci, "Missing link TRB at end of ring segment\n");
226 --- a/drivers/usb/host/xhci.c
227 +++ b/drivers/usb/host/xhci.c
228 @@ -895,8 +895,8 @@ static void xhci_clear_command_ring(stru
232 - sizeof(union xhci_trb) * (TRBS_PER_SEGMENT - 1));
233 - seg->trbs[TRBS_PER_SEGMENT - 1].link.control &=
234 + sizeof(union xhci_trb) * (ring->trbs_per_seg - 1));
235 + seg->trbs[ring->trbs_per_seg - 1].link.control &=
236 cpu_to_le32(~TRB_CYCLE);
238 } while (seg != ring->deq_seg);
239 @@ -907,7 +907,7 @@ static void xhci_clear_command_ring(stru
240 ring->enq_seg = ring->deq_seg;
241 ring->enqueue = ring->dequeue;
243 - ring->num_trbs_free = ring->num_segs * (TRBS_PER_SEGMENT - 1) - 1;
244 + ring->num_trbs_free = ring->num_segs * (ring->trbs_per_seg - 1) - 1;
246 * Ring is now zeroed, so the HW should look for change of ownership
247 * when the cycle bit is set to 1.
248 --- a/drivers/usb/host/xhci.h
249 +++ b/drivers/usb/host/xhci.h
250 @@ -1634,6 +1634,7 @@ struct xhci_ring {
251 unsigned int num_trbs_free;
252 unsigned int num_trbs_free_temp;
253 unsigned int bounce_buf_len;
254 + unsigned int trbs_per_seg;
255 enum xhci_ring_type type;
256 bool last_td_was_short;
257 struct radix_tree_root *trb_address_map;