1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * SPDX-License-Identifier: MIT
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * Copyright © 2019 Intel Corporation
5*4882a593Smuzhiyun */
6*4882a593Smuzhiyun
7*4882a593Smuzhiyun #ifndef INTEL_RING_H
8*4882a593Smuzhiyun #define INTEL_RING_H
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #include "i915_gem.h" /* GEM_BUG_ON */
11*4882a593Smuzhiyun #include "i915_request.h"
12*4882a593Smuzhiyun #include "intel_ring_types.h"
13*4882a593Smuzhiyun
14*4882a593Smuzhiyun struct intel_engine_cs;
15*4882a593Smuzhiyun
16*4882a593Smuzhiyun struct intel_ring *
17*4882a593Smuzhiyun intel_engine_create_ring(struct intel_engine_cs *engine, int size);
18*4882a593Smuzhiyun
19*4882a593Smuzhiyun u32 *intel_ring_begin(struct i915_request *rq, unsigned int num_dwords);
20*4882a593Smuzhiyun int intel_ring_cacheline_align(struct i915_request *rq);
21*4882a593Smuzhiyun
22*4882a593Smuzhiyun unsigned int intel_ring_update_space(struct intel_ring *ring);
23*4882a593Smuzhiyun
24*4882a593Smuzhiyun void __intel_ring_pin(struct intel_ring *ring);
25*4882a593Smuzhiyun int intel_ring_pin(struct intel_ring *ring, struct i915_gem_ww_ctx *ww);
26*4882a593Smuzhiyun void intel_ring_unpin(struct intel_ring *ring);
27*4882a593Smuzhiyun void intel_ring_reset(struct intel_ring *ring, u32 tail);
28*4882a593Smuzhiyun
29*4882a593Smuzhiyun void intel_ring_free(struct kref *ref);
30*4882a593Smuzhiyun
intel_ring_get(struct intel_ring * ring)31*4882a593Smuzhiyun static inline struct intel_ring *intel_ring_get(struct intel_ring *ring)
32*4882a593Smuzhiyun {
33*4882a593Smuzhiyun kref_get(&ring->ref);
34*4882a593Smuzhiyun return ring;
35*4882a593Smuzhiyun }
36*4882a593Smuzhiyun
intel_ring_put(struct intel_ring * ring)37*4882a593Smuzhiyun static inline void intel_ring_put(struct intel_ring *ring)
38*4882a593Smuzhiyun {
39*4882a593Smuzhiyun kref_put(&ring->ref, intel_ring_free);
40*4882a593Smuzhiyun }
41*4882a593Smuzhiyun
intel_ring_advance(struct i915_request * rq,u32 * cs)42*4882a593Smuzhiyun static inline void intel_ring_advance(struct i915_request *rq, u32 *cs)
43*4882a593Smuzhiyun {
44*4882a593Smuzhiyun /* Dummy function.
45*4882a593Smuzhiyun *
46*4882a593Smuzhiyun * This serves as a placeholder in the code so that the reader
47*4882a593Smuzhiyun * can compare against the preceding intel_ring_begin() and
48*4882a593Smuzhiyun * check that the number of dwords emitted matches the space
49*4882a593Smuzhiyun * reserved for the command packet (i.e. the value passed to
50*4882a593Smuzhiyun * intel_ring_begin()).
51*4882a593Smuzhiyun */
52*4882a593Smuzhiyun GEM_BUG_ON((rq->ring->vaddr + rq->ring->emit) != cs);
53*4882a593Smuzhiyun }
54*4882a593Smuzhiyun
intel_ring_wrap(const struct intel_ring * ring,u32 pos)55*4882a593Smuzhiyun static inline u32 intel_ring_wrap(const struct intel_ring *ring, u32 pos)
56*4882a593Smuzhiyun {
57*4882a593Smuzhiyun return pos & (ring->size - 1);
58*4882a593Smuzhiyun }
59*4882a593Smuzhiyun
intel_ring_direction(const struct intel_ring * ring,u32 next,u32 prev)60*4882a593Smuzhiyun static inline int intel_ring_direction(const struct intel_ring *ring,
61*4882a593Smuzhiyun u32 next, u32 prev)
62*4882a593Smuzhiyun {
63*4882a593Smuzhiyun typecheck(typeof(ring->size), next);
64*4882a593Smuzhiyun typecheck(typeof(ring->size), prev);
65*4882a593Smuzhiyun return (next - prev) << ring->wrap;
66*4882a593Smuzhiyun }
67*4882a593Smuzhiyun
68*4882a593Smuzhiyun static inline bool
intel_ring_offset_valid(const struct intel_ring * ring,unsigned int pos)69*4882a593Smuzhiyun intel_ring_offset_valid(const struct intel_ring *ring,
70*4882a593Smuzhiyun unsigned int pos)
71*4882a593Smuzhiyun {
72*4882a593Smuzhiyun if (pos & -ring->size) /* must be strictly within the ring */
73*4882a593Smuzhiyun return false;
74*4882a593Smuzhiyun
75*4882a593Smuzhiyun if (!IS_ALIGNED(pos, 8)) /* must be qword aligned */
76*4882a593Smuzhiyun return false;
77*4882a593Smuzhiyun
78*4882a593Smuzhiyun return true;
79*4882a593Smuzhiyun }
80*4882a593Smuzhiyun
intel_ring_offset(const struct i915_request * rq,void * addr)81*4882a593Smuzhiyun static inline u32 intel_ring_offset(const struct i915_request *rq, void *addr)
82*4882a593Smuzhiyun {
83*4882a593Smuzhiyun /* Don't write ring->size (equivalent to 0) as that hangs some GPUs. */
84*4882a593Smuzhiyun u32 offset = addr - rq->ring->vaddr;
85*4882a593Smuzhiyun GEM_BUG_ON(offset > rq->ring->size);
86*4882a593Smuzhiyun return intel_ring_wrap(rq->ring, offset);
87*4882a593Smuzhiyun }
88*4882a593Smuzhiyun
89*4882a593Smuzhiyun static inline void
assert_ring_tail_valid(const struct intel_ring * ring,unsigned int tail)90*4882a593Smuzhiyun assert_ring_tail_valid(const struct intel_ring *ring, unsigned int tail)
91*4882a593Smuzhiyun {
92*4882a593Smuzhiyun unsigned int head = READ_ONCE(ring->head);
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun GEM_BUG_ON(!intel_ring_offset_valid(ring, tail));
95*4882a593Smuzhiyun
96*4882a593Smuzhiyun /*
97*4882a593Smuzhiyun * "Ring Buffer Use"
98*4882a593Smuzhiyun * Gen2 BSpec "1. Programming Environment" / 1.4.4.6
99*4882a593Smuzhiyun * Gen3 BSpec "1c Memory Interface Functions" / 2.3.4.5
100*4882a593Smuzhiyun * Gen4+ BSpec "1c Memory Interface and Command Stream" / 5.3.4.5
101*4882a593Smuzhiyun * "If the Ring Buffer Head Pointer and the Tail Pointer are on the
102*4882a593Smuzhiyun * same cacheline, the Head Pointer must not be greater than the Tail
103*4882a593Smuzhiyun * Pointer."
104*4882a593Smuzhiyun *
105*4882a593Smuzhiyun * We use ring->head as the last known location of the actual RING_HEAD,
106*4882a593Smuzhiyun * it may have advanced but in the worst case it is equally the same
107*4882a593Smuzhiyun * as ring->head and so we should never program RING_TAIL to advance
108*4882a593Smuzhiyun * into the same cacheline as ring->head.
109*4882a593Smuzhiyun */
110*4882a593Smuzhiyun #define cacheline(a) round_down(a, CACHELINE_BYTES)
111*4882a593Smuzhiyun GEM_BUG_ON(cacheline(tail) == cacheline(head) && tail < head);
112*4882a593Smuzhiyun #undef cacheline
113*4882a593Smuzhiyun }
114*4882a593Smuzhiyun
115*4882a593Smuzhiyun static inline unsigned int
intel_ring_set_tail(struct intel_ring * ring,unsigned int tail)116*4882a593Smuzhiyun intel_ring_set_tail(struct intel_ring *ring, unsigned int tail)
117*4882a593Smuzhiyun {
118*4882a593Smuzhiyun /* Whilst writes to the tail are strictly order, there is no
119*4882a593Smuzhiyun * serialisation between readers and the writers. The tail may be
120*4882a593Smuzhiyun * read by i915_request_retire() just as it is being updated
121*4882a593Smuzhiyun * by execlists, as although the breadcrumb is complete, the context
122*4882a593Smuzhiyun * switch hasn't been seen.
123*4882a593Smuzhiyun */
124*4882a593Smuzhiyun assert_ring_tail_valid(ring, tail);
125*4882a593Smuzhiyun ring->tail = tail;
126*4882a593Smuzhiyun return tail;
127*4882a593Smuzhiyun }
128*4882a593Smuzhiyun
129*4882a593Smuzhiyun static inline unsigned int
__intel_ring_space(unsigned int head,unsigned int tail,unsigned int size)130*4882a593Smuzhiyun __intel_ring_space(unsigned int head, unsigned int tail, unsigned int size)
131*4882a593Smuzhiyun {
132*4882a593Smuzhiyun /*
133*4882a593Smuzhiyun * "If the Ring Buffer Head Pointer and the Tail Pointer are on the
134*4882a593Smuzhiyun * same cacheline, the Head Pointer must not be greater than the Tail
135*4882a593Smuzhiyun * Pointer."
136*4882a593Smuzhiyun */
137*4882a593Smuzhiyun GEM_BUG_ON(!is_power_of_2(size));
138*4882a593Smuzhiyun return (head - tail - CACHELINE_BYTES) & (size - 1);
139*4882a593Smuzhiyun }
140*4882a593Smuzhiyun
141*4882a593Smuzhiyun #endif /* INTEL_RING_H */
142