1fb4d8502Sjsg /*
2fb4d8502Sjsg * Copyright 2015 Advanced Micro Devices, Inc.
3fb4d8502Sjsg *
4fb4d8502Sjsg * Permission is hereby granted, free of charge, to any person obtaining a
5fb4d8502Sjsg * copy of this software and associated documentation files (the "Software"),
6fb4d8502Sjsg * to deal in the Software without restriction, including without limitation
7fb4d8502Sjsg * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8fb4d8502Sjsg * and/or sell copies of the Software, and to permit persons to whom the
9fb4d8502Sjsg * Software is furnished to do so, subject to the following conditions:
10fb4d8502Sjsg *
11fb4d8502Sjsg * The above copyright notice and this permission notice shall be included in
12fb4d8502Sjsg * all copies or substantial portions of the Software.
13fb4d8502Sjsg *
14fb4d8502Sjsg * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15fb4d8502Sjsg * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16fb4d8502Sjsg * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17fb4d8502Sjsg * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18fb4d8502Sjsg * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19fb4d8502Sjsg * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20fb4d8502Sjsg * OTHER DEALINGS IN THE SOFTWARE.
21fb4d8502Sjsg *
22fb4d8502Sjsg */
23c349dbc7Sjsg
24c349dbc7Sjsg #include <linux/pci.h>
25c349dbc7Sjsg
26fb4d8502Sjsg #include "amdgpu.h"
27fb4d8502Sjsg #include "amdgpu_ih.h"
28fb4d8502Sjsg #include "sid.h"
29fb4d8502Sjsg #include "si_ih.h"
30ad8b1aafSjsg #include "oss/oss_1_0_d.h"
31ad8b1aafSjsg #include "oss/oss_1_0_sh_mask.h"
32fb4d8502Sjsg
33fb4d8502Sjsg static void si_ih_set_interrupt_funcs(struct amdgpu_device *adev);
34fb4d8502Sjsg
si_ih_enable_interrupts(struct amdgpu_device * adev)35fb4d8502Sjsg static void si_ih_enable_interrupts(struct amdgpu_device *adev)
36fb4d8502Sjsg {
37fb4d8502Sjsg u32 ih_cntl = RREG32(IH_CNTL);
38fb4d8502Sjsg u32 ih_rb_cntl = RREG32(IH_RB_CNTL);
39fb4d8502Sjsg
40fb4d8502Sjsg ih_cntl |= ENABLE_INTR;
41fb4d8502Sjsg ih_rb_cntl |= IH_RB_ENABLE;
42fb4d8502Sjsg WREG32(IH_CNTL, ih_cntl);
43fb4d8502Sjsg WREG32(IH_RB_CNTL, ih_rb_cntl);
44fb4d8502Sjsg adev->irq.ih.enabled = true;
45fb4d8502Sjsg }
46fb4d8502Sjsg
si_ih_disable_interrupts(struct amdgpu_device * adev)47fb4d8502Sjsg static void si_ih_disable_interrupts(struct amdgpu_device *adev)
48fb4d8502Sjsg {
49fb4d8502Sjsg u32 ih_rb_cntl = RREG32(IH_RB_CNTL);
50fb4d8502Sjsg u32 ih_cntl = RREG32(IH_CNTL);
51fb4d8502Sjsg
52fb4d8502Sjsg ih_rb_cntl &= ~IH_RB_ENABLE;
53fb4d8502Sjsg ih_cntl &= ~ENABLE_INTR;
54fb4d8502Sjsg WREG32(IH_RB_CNTL, ih_rb_cntl);
55fb4d8502Sjsg WREG32(IH_CNTL, ih_cntl);
56fb4d8502Sjsg WREG32(IH_RB_RPTR, 0);
57fb4d8502Sjsg WREG32(IH_RB_WPTR, 0);
58fb4d8502Sjsg adev->irq.ih.enabled = false;
59fb4d8502Sjsg adev->irq.ih.rptr = 0;
60fb4d8502Sjsg }
61fb4d8502Sjsg
si_ih_irq_init(struct amdgpu_device * adev)62fb4d8502Sjsg static int si_ih_irq_init(struct amdgpu_device *adev)
63fb4d8502Sjsg {
64c349dbc7Sjsg struct amdgpu_ih_ring *ih = &adev->irq.ih;
65fb4d8502Sjsg int rb_bufsz;
66fb4d8502Sjsg u32 interrupt_cntl, ih_cntl, ih_rb_cntl;
67fb4d8502Sjsg
68fb4d8502Sjsg si_ih_disable_interrupts(adev);
6938f65c9fSjsg /* set dummy read address to dummy page address */
7038f65c9fSjsg WREG32(INTERRUPT_CNTL2, adev->dummy_page_addr >> 8);
71fb4d8502Sjsg interrupt_cntl = RREG32(INTERRUPT_CNTL);
72fb4d8502Sjsg interrupt_cntl &= ~IH_DUMMY_RD_OVERRIDE;
73fb4d8502Sjsg interrupt_cntl &= ~IH_REQ_NONSNOOP_EN;
74fb4d8502Sjsg WREG32(INTERRUPT_CNTL, interrupt_cntl);
75fb4d8502Sjsg
76fb4d8502Sjsg WREG32(IH_RB_BASE, adev->irq.ih.gpu_addr >> 8);
77fb4d8502Sjsg rb_bufsz = order_base_2(adev->irq.ih.ring_size / 4);
78fb4d8502Sjsg
79fb4d8502Sjsg ih_rb_cntl = IH_WPTR_OVERFLOW_ENABLE |
80fb4d8502Sjsg IH_WPTR_OVERFLOW_CLEAR |
81fb4d8502Sjsg (rb_bufsz << 1) |
82fb4d8502Sjsg IH_WPTR_WRITEBACK_ENABLE;
83fb4d8502Sjsg
84c349dbc7Sjsg WREG32(IH_RB_WPTR_ADDR_LO, lower_32_bits(ih->wptr_addr));
85c349dbc7Sjsg WREG32(IH_RB_WPTR_ADDR_HI, upper_32_bits(ih->wptr_addr) & 0xFF);
86fb4d8502Sjsg WREG32(IH_RB_CNTL, ih_rb_cntl);
87fb4d8502Sjsg WREG32(IH_RB_RPTR, 0);
88fb4d8502Sjsg WREG32(IH_RB_WPTR, 0);
89fb4d8502Sjsg
90fb4d8502Sjsg ih_cntl = MC_WRREQ_CREDIT(0x10) | MC_WR_CLEAN_CNT(0x10) | MC_VMID(0);
91fb4d8502Sjsg if (adev->irq.msi_enabled)
92fb4d8502Sjsg ih_cntl |= RPTR_REARM;
93fb4d8502Sjsg WREG32(IH_CNTL, ih_cntl);
94fb4d8502Sjsg
95fb4d8502Sjsg pci_set_master(adev->pdev);
96fb4d8502Sjsg si_ih_enable_interrupts(adev);
97fb4d8502Sjsg
98fb4d8502Sjsg return 0;
99fb4d8502Sjsg }
100fb4d8502Sjsg
si_ih_irq_disable(struct amdgpu_device * adev)101fb4d8502Sjsg static void si_ih_irq_disable(struct amdgpu_device *adev)
102fb4d8502Sjsg {
103fb4d8502Sjsg si_ih_disable_interrupts(adev);
104fb4d8502Sjsg mdelay(1);
105fb4d8502Sjsg }
106fb4d8502Sjsg
si_ih_get_wptr(struct amdgpu_device * adev,struct amdgpu_ih_ring * ih)107c349dbc7Sjsg static u32 si_ih_get_wptr(struct amdgpu_device *adev,
108c349dbc7Sjsg struct amdgpu_ih_ring *ih)
109fb4d8502Sjsg {
110fb4d8502Sjsg u32 wptr, tmp;
111fb4d8502Sjsg
112c349dbc7Sjsg wptr = le32_to_cpu(*ih->wptr_cpu);
113fb4d8502Sjsg
114fb4d8502Sjsg if (wptr & IH_RB_WPTR__RB_OVERFLOW_MASK) {
115fb4d8502Sjsg wptr &= ~IH_RB_WPTR__RB_OVERFLOW_MASK;
116fb4d8502Sjsg dev_warn(adev->dev, "IH ring buffer overflow (0x%08X, 0x%08X, 0x%08X)\n",
117c349dbc7Sjsg wptr, ih->rptr, (wptr + 16) & ih->ptr_mask);
118c349dbc7Sjsg ih->rptr = (wptr + 16) & ih->ptr_mask;
119fb4d8502Sjsg tmp = RREG32(IH_RB_CNTL);
120fb4d8502Sjsg tmp |= IH_RB_CNTL__WPTR_OVERFLOW_CLEAR_MASK;
121fb4d8502Sjsg WREG32(IH_RB_CNTL, tmp);
122*f7a572c7Sjsg
123*f7a572c7Sjsg /* Unset the CLEAR_OVERFLOW bit immediately so new overflows
124*f7a572c7Sjsg * can be detected.
125*f7a572c7Sjsg */
126*f7a572c7Sjsg tmp &= ~IH_RB_CNTL__WPTR_OVERFLOW_CLEAR_MASK;
127*f7a572c7Sjsg WREG32(IH_RB_CNTL, tmp);
128fb4d8502Sjsg }
129c349dbc7Sjsg return (wptr & ih->ptr_mask);
130fb4d8502Sjsg }
131fb4d8502Sjsg
si_ih_decode_iv(struct amdgpu_device * adev,struct amdgpu_ih_ring * ih,struct amdgpu_iv_entry * entry)132fb4d8502Sjsg static void si_ih_decode_iv(struct amdgpu_device *adev,
133c349dbc7Sjsg struct amdgpu_ih_ring *ih,
134fb4d8502Sjsg struct amdgpu_iv_entry *entry)
135fb4d8502Sjsg {
136c349dbc7Sjsg u32 ring_index = ih->rptr >> 2;
137fb4d8502Sjsg uint32_t dw[4];
138fb4d8502Sjsg
139c349dbc7Sjsg dw[0] = le32_to_cpu(ih->ring[ring_index + 0]);
140c349dbc7Sjsg dw[1] = le32_to_cpu(ih->ring[ring_index + 1]);
141c349dbc7Sjsg dw[2] = le32_to_cpu(ih->ring[ring_index + 2]);
142c349dbc7Sjsg dw[3] = le32_to_cpu(ih->ring[ring_index + 3]);
143fb4d8502Sjsg
144c349dbc7Sjsg entry->client_id = AMDGPU_IRQ_CLIENTID_LEGACY;
145fb4d8502Sjsg entry->src_id = dw[0] & 0xff;
146fb4d8502Sjsg entry->src_data[0] = dw[1] & 0xfffffff;
147fb4d8502Sjsg entry->ring_id = dw[2] & 0xff;
148fb4d8502Sjsg entry->vmid = (dw[2] >> 8) & 0xff;
149fb4d8502Sjsg
150c349dbc7Sjsg ih->rptr += 16;
151fb4d8502Sjsg }
152fb4d8502Sjsg
si_ih_set_rptr(struct amdgpu_device * adev,struct amdgpu_ih_ring * ih)153c349dbc7Sjsg static void si_ih_set_rptr(struct amdgpu_device *adev,
154c349dbc7Sjsg struct amdgpu_ih_ring *ih)
155fb4d8502Sjsg {
156c349dbc7Sjsg WREG32(IH_RB_RPTR, ih->rptr);
157fb4d8502Sjsg }
158fb4d8502Sjsg
si_ih_early_init(void * handle)159fb4d8502Sjsg static int si_ih_early_init(void *handle)
160fb4d8502Sjsg {
161fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
162fb4d8502Sjsg
163fb4d8502Sjsg si_ih_set_interrupt_funcs(adev);
164fb4d8502Sjsg
165fb4d8502Sjsg return 0;
166fb4d8502Sjsg }
167fb4d8502Sjsg
si_ih_sw_init(void * handle)168fb4d8502Sjsg static int si_ih_sw_init(void *handle)
169fb4d8502Sjsg {
170fb4d8502Sjsg int r;
171fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
172fb4d8502Sjsg
173c349dbc7Sjsg r = amdgpu_ih_ring_init(adev, &adev->irq.ih, 64 * 1024, false);
174fb4d8502Sjsg if (r)
175fb4d8502Sjsg return r;
176fb4d8502Sjsg
177fb4d8502Sjsg return amdgpu_irq_init(adev);
178fb4d8502Sjsg }
179fb4d8502Sjsg
si_ih_sw_fini(void * handle)180fb4d8502Sjsg static int si_ih_sw_fini(void *handle)
181fb4d8502Sjsg {
182fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
183fb4d8502Sjsg
1845ca02815Sjsg amdgpu_irq_fini_sw(adev);
185fb4d8502Sjsg
186fb4d8502Sjsg return 0;
187fb4d8502Sjsg }
188fb4d8502Sjsg
si_ih_hw_init(void * handle)189fb4d8502Sjsg static int si_ih_hw_init(void *handle)
190fb4d8502Sjsg {
191fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
192fb4d8502Sjsg
193fb4d8502Sjsg return si_ih_irq_init(adev);
194fb4d8502Sjsg }
195fb4d8502Sjsg
si_ih_hw_fini(void * handle)196fb4d8502Sjsg static int si_ih_hw_fini(void *handle)
197fb4d8502Sjsg {
198fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
199fb4d8502Sjsg
200fb4d8502Sjsg si_ih_irq_disable(adev);
201fb4d8502Sjsg
202fb4d8502Sjsg return 0;
203fb4d8502Sjsg }
204fb4d8502Sjsg
si_ih_suspend(void * handle)205fb4d8502Sjsg static int si_ih_suspend(void *handle)
206fb4d8502Sjsg {
207fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
208fb4d8502Sjsg
209fb4d8502Sjsg return si_ih_hw_fini(adev);
210fb4d8502Sjsg }
211fb4d8502Sjsg
si_ih_resume(void * handle)212fb4d8502Sjsg static int si_ih_resume(void *handle)
213fb4d8502Sjsg {
214fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
215fb4d8502Sjsg
216fb4d8502Sjsg return si_ih_hw_init(adev);
217fb4d8502Sjsg }
218fb4d8502Sjsg
si_ih_is_idle(void * handle)219fb4d8502Sjsg static bool si_ih_is_idle(void *handle)
220fb4d8502Sjsg {
221fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
222fb4d8502Sjsg u32 tmp = RREG32(SRBM_STATUS);
223fb4d8502Sjsg
224fb4d8502Sjsg if (tmp & SRBM_STATUS__IH_BUSY_MASK)
225fb4d8502Sjsg return false;
226fb4d8502Sjsg
227fb4d8502Sjsg return true;
228fb4d8502Sjsg }
229fb4d8502Sjsg
si_ih_wait_for_idle(void * handle)230fb4d8502Sjsg static int si_ih_wait_for_idle(void *handle)
231fb4d8502Sjsg {
232fb4d8502Sjsg unsigned i;
233fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
234fb4d8502Sjsg
235fb4d8502Sjsg for (i = 0; i < adev->usec_timeout; i++) {
236fb4d8502Sjsg if (si_ih_is_idle(handle))
237fb4d8502Sjsg return 0;
238fb4d8502Sjsg udelay(1);
239fb4d8502Sjsg }
240fb4d8502Sjsg return -ETIMEDOUT;
241fb4d8502Sjsg }
242fb4d8502Sjsg
si_ih_soft_reset(void * handle)243fb4d8502Sjsg static int si_ih_soft_reset(void *handle)
244fb4d8502Sjsg {
245fb4d8502Sjsg struct amdgpu_device *adev = (struct amdgpu_device *)handle;
246fb4d8502Sjsg
247fb4d8502Sjsg u32 srbm_soft_reset = 0;
248fb4d8502Sjsg u32 tmp = RREG32(SRBM_STATUS);
249fb4d8502Sjsg
250fb4d8502Sjsg if (tmp & SRBM_STATUS__IH_BUSY_MASK)
251fb4d8502Sjsg srbm_soft_reset |= SRBM_SOFT_RESET__SOFT_RESET_IH_MASK;
252fb4d8502Sjsg
253fb4d8502Sjsg if (srbm_soft_reset) {
254fb4d8502Sjsg tmp = RREG32(SRBM_SOFT_RESET);
255fb4d8502Sjsg tmp |= srbm_soft_reset;
256fb4d8502Sjsg dev_info(adev->dev, "SRBM_SOFT_RESET=0x%08X\n", tmp);
257fb4d8502Sjsg WREG32(SRBM_SOFT_RESET, tmp);
258fb4d8502Sjsg tmp = RREG32(SRBM_SOFT_RESET);
259fb4d8502Sjsg
260fb4d8502Sjsg udelay(50);
261fb4d8502Sjsg
262fb4d8502Sjsg tmp &= ~srbm_soft_reset;
263fb4d8502Sjsg WREG32(SRBM_SOFT_RESET, tmp);
264fb4d8502Sjsg tmp = RREG32(SRBM_SOFT_RESET);
265fb4d8502Sjsg
266fb4d8502Sjsg udelay(50);
267fb4d8502Sjsg }
268fb4d8502Sjsg
269fb4d8502Sjsg return 0;
270fb4d8502Sjsg }
271fb4d8502Sjsg
si_ih_set_clockgating_state(void * handle,enum amd_clockgating_state state)272fb4d8502Sjsg static int si_ih_set_clockgating_state(void *handle,
273fb4d8502Sjsg enum amd_clockgating_state state)
274fb4d8502Sjsg {
275fb4d8502Sjsg return 0;
276fb4d8502Sjsg }
277fb4d8502Sjsg
si_ih_set_powergating_state(void * handle,enum amd_powergating_state state)278fb4d8502Sjsg static int si_ih_set_powergating_state(void *handle,
279fb4d8502Sjsg enum amd_powergating_state state)
280fb4d8502Sjsg {
281fb4d8502Sjsg return 0;
282fb4d8502Sjsg }
283fb4d8502Sjsg
284fb4d8502Sjsg static const struct amd_ip_funcs si_ih_ip_funcs = {
285fb4d8502Sjsg .name = "si_ih",
286fb4d8502Sjsg .early_init = si_ih_early_init,
287fb4d8502Sjsg .late_init = NULL,
288fb4d8502Sjsg .sw_init = si_ih_sw_init,
289fb4d8502Sjsg .sw_fini = si_ih_sw_fini,
290fb4d8502Sjsg .hw_init = si_ih_hw_init,
291fb4d8502Sjsg .hw_fini = si_ih_hw_fini,
292fb4d8502Sjsg .suspend = si_ih_suspend,
293fb4d8502Sjsg .resume = si_ih_resume,
294fb4d8502Sjsg .is_idle = si_ih_is_idle,
295fb4d8502Sjsg .wait_for_idle = si_ih_wait_for_idle,
296fb4d8502Sjsg .soft_reset = si_ih_soft_reset,
297fb4d8502Sjsg .set_clockgating_state = si_ih_set_clockgating_state,
298fb4d8502Sjsg .set_powergating_state = si_ih_set_powergating_state,
299fb4d8502Sjsg };
300fb4d8502Sjsg
301fb4d8502Sjsg static const struct amdgpu_ih_funcs si_ih_funcs = {
302fb4d8502Sjsg .get_wptr = si_ih_get_wptr,
303fb4d8502Sjsg .decode_iv = si_ih_decode_iv,
304fb4d8502Sjsg .set_rptr = si_ih_set_rptr
305fb4d8502Sjsg };
306fb4d8502Sjsg
si_ih_set_interrupt_funcs(struct amdgpu_device * adev)307fb4d8502Sjsg static void si_ih_set_interrupt_funcs(struct amdgpu_device *adev)
308fb4d8502Sjsg {
309fb4d8502Sjsg adev->irq.ih_funcs = &si_ih_funcs;
310fb4d8502Sjsg }
311fb4d8502Sjsg
312fb4d8502Sjsg const struct amdgpu_ip_block_version si_ih_ip_block =
313fb4d8502Sjsg {
314fb4d8502Sjsg .type = AMD_IP_BLOCK_TYPE_IH,
315fb4d8502Sjsg .major = 1,
316fb4d8502Sjsg .minor = 0,
317fb4d8502Sjsg .rev = 0,
318fb4d8502Sjsg .funcs = &si_ih_ip_funcs,
319fb4d8502Sjsg };
320