xref: /netbsd-src/sys/arch/sparc64/include/hypervisor.h (revision a24efa7dea9f1f56c3bdb15a927d3516792ace1c)
1 /*	$NetBSD: hypervisor.h,v 1.5 2014/09/24 18:32:10 palle Exp $ */
2 /*	$OpenBSD: hypervisor.h,v 1.14 2011/06/26 17:23:46 kettenis Exp $	*/
3 
4 /*
5  * Copyright (c) 2008 Mark Kettenis
6  *
7  * Permission to use, copy, modify, and distribute this software for any
8  * purpose with or without fee is hereby granted, provided that the above
9  * copyright notice and this permission notice appear in all copies.
10  *
11  * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
12  * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
13  * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
14  * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
15  * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
16  * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
17  * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
18  */
19 
20 #ifndef	_HYPERVISOR_H_
21 #define	_HYPERVISOR_H_
22 
23 /*
24  * UltraSPARC Hypervisor API.
25  */
26 
27 /*
28  * FAST_TRAP function numbers
29  */
30 
31 #define FT_MMU_MAP_PERM_ADDR 0x25
32 
33 /*
34  * API versioning
35  */
36 
37 #ifndef _LOCORE
38 int64_t	hv_api_get_version(uint64_t api_group,
39 	    uint64_t *major_number, uint64_t *minor_number);
40 #endif
41 /*
42  * Domain services
43  */
44 
45 #ifndef _LOCORE
46 int64_t hv_mach_desc(paddr_t buffer, psize_t *length);
47 #endif
48 
49 /*
50  * CPU services
51  */
52 
53 #ifndef _LOCORE
54 void	hv_cpu_yield(void);
55 int64_t	hv_cpu_qconf(uint64_t queue, uint64_t base, uint64_t nentries);
56 #endif
57 
58 #define CPU_MONDO_QUEUE		0x3c
59 #define DEVICE_MONDO_QUEUE	0x3d
60 
61 #ifndef _LOCORE
62 int64_t	hv_cpu_mondo_send(uint64_t ncpus, paddr_t cpulist, paddr_t data);
63 int64_t	hv_cpu_myid(uint64_t *cpuid);
64 #endif
65 
66 /*
67  * MMU services
68  */
69 
70 #ifndef _LOCORE
71 int64_t	hv_mmu_demap_page(vaddr_t vaddr, uint64_t context, uint64_t flags);
72 int64_t	hv_mmu_demap_ctx(uint64_t context, uint64_t flags);
73 int64_t	hv_mmu_demap_all(uint64_t flags);
74 int64_t	hv_mmu_map_perm_addr(vaddr_t vaddr, uint64_t tte, uint64_t flags);
75 int64_t	hv_mmu_unmap_perm_addr(vaddr_t vaddr, uint64_t flags);
76 int64_t	hv_mmu_map_addr(vaddr_t vaddr, uint64_t context, uint64_t tte,
77 	    uint64_t flags);
78 int64_t	hv_mmu_unmap_addr(vaddr_t vaddr, uint64_t context, uint64_t flags);
79 #endif
80 
81 #define MAP_DTLB	0x1
82 #define MAP_ITLB	0x2
83 
84 #ifndef _LOCORE
85 struct tsb_desc {
86 	uint16_t	td_idxpgsz;
87 	uint16_t	td_assoc;
88 	uint32_t	td_size;
89 	uint32_t	td_ctxidx;
90 	uint32_t	td_pgsz;
91 	paddr_t		td_pa;
92 	uint64_t	td_reserved;
93 };
94 
95 int64_t	hv_mmu_tsb_ctx0(uint64_t ntsb, paddr_t tsbptr);
96 int64_t	hv_mmu_tsb_ctxnon0(uint64_t ntsb, paddr_t tsbptr);
97 #endif
98 
99 /*
100  * Cache and memory services
101  */
102 
103 #ifndef _LOCORE
104 int64_t	hv_mem_scrub(paddr_t raddr, psize_t length);
105 int64_t	hv_mem_sync(paddr_t raddr, psize_t length);
106 #endif
107 
108 /*
109  * Device interrupt services
110  */
111 
112 #ifndef _LOCORE
113 int64_t	hv_intr_devino_to_sysino(uint64_t devhandle, uint64_t devino,
114 	    uint64_t *sysino);
115 int64_t	hv_intr_getenabled(uint64_t sysino, uint64_t *intr_enabled);
116 int64_t	hv_intr_setenabled(uint64_t sysino, uint64_t intr_enabled);
117 int64_t	hv_intr_getstate(uint64_t sysino, uint64_t *intr_state);
118 int64_t	hv_intr_setstate(uint64_t sysino, uint64_t intr_state);
119 int64_t	hv_intr_gettarget(uint64_t sysino, uint64_t *cpuid);
120 int64_t	hv_intr_settarget(uint64_t sysino, uint64_t cpuid);
121 #endif
122 
123 #define INTR_DISABLED	0
124 #define INTR_ENABLED	1
125 
126 #define INTR_IDLE	0
127 #define INTR_RECEIVED	1
128 #define INTR_DELIVERED	2
129 
130 #ifndef _LOCORE
131 int64_t	hv_vintr_getcookie(uint64_t devhandle, uint64_t devino,
132 	    uint64_t *cookie_value);
133 int64_t	hv_vintr_setcookie(uint64_t devhandle, uint64_t devino,
134 	    uint64_t cookie_value);
135 int64_t	hv_vintr_getenabled(uint64_t devhandle, uint64_t devino,
136 	    uint64_t *intr_enabled);
137 int64_t	hv_vintr_setenabled(uint64_t devhandle, uint64_t devino,
138 	    uint64_t intr_enabled);
139 int64_t	hv_vintr_getstate(uint64_t devhandle, uint64_t devino,
140 	    uint64_t *intr_state);
141 int64_t	hv_vintr_setstate(uint64_t devhandle, uint64_t devino,
142 	    uint64_t intr_state);
143 int64_t	hv_vintr_gettarget(uint64_t devhandle, uint64_t devino,
144 	    uint64_t *cpuid);
145 int64_t	hv_vintr_settarget(uint64_t devhandle, uint64_t devino,
146 	    uint64_t cpuid);
147 #endif
148 
149 /*
150  * Time of day services
151  */
152 
153 #ifndef _LOCORE
154 int64_t	hv_tod_get(uint64_t *tod);
155 int64_t	hv_tod_set(uint64_t tod);
156 #endif
157 
158 /*
159  * Console services
160  */
161 
162 #ifndef _LOCORE
163 int64_t	hv_cons_getchar(int64_t *ch);
164 int64_t	hv_cons_putchar(int64_t ch);
165 int64_t	hv_api_putchar(int64_t ch);
166 #endif
167 
168 #define CONS_BREAK	-1
169 #define CONS_HUP	-2
170 
171 /*
172  * Domain state services
173  */
174 
175 #ifndef _LOCORE
176 int64_t	hv_soft_state_set(uint64_t software_state,
177 	    paddr_t software_description_ptr);
178 #endif
179 
180 #define SIS_NORMAL	0x1
181 #define SIS_TRANSITION	0x2
182 
183 /*
184  * PCI I/O services
185  */
186 
187 #ifndef _LOCORE
188 int64_t	hv_pci_iommu_map(uint64_t devhandle, uint64_t tsbid,
189 	    uint64_t nttes, uint64_t io_attributes, paddr_t io_page_list_p,
190 	    uint64_t *nttes_mapped);
191 int64_t	hv_pci_iommu_demap(uint64_t devhandle, uint64_t tsbid,
192 	    uint64_t nttes, uint64_t *nttes_demapped);
193 int64_t	hv_pci_iommu_getmap(uint64_t devhandle, uint64_t tsbid,
194 	    uint64_t *io_attributes, paddr_t *r_addr);
195 int64_t	hv_pci_iommu_getbypass(uint64_t devhandle, paddr_t r_addr,
196 	    uint64_t io_attributes, uint64_t *io_addr);
197 
198 int64_t	hv_pci_config_get(uint64_t devhandle, uint64_t pci_device,
199             uint64_t pci_config_offset, uint64_t size,
200 	    uint64_t *error_flag, uint64_t *data);
201 int64_t	hv_pci_config_put(uint64_t devhandle, uint64_t pci_device,
202             uint64_t pci_config_offset, uint64_t size, uint64_t data,
203 	    uint64_t *error_flag);
204 #endif
205 
206 #define PCI_MAP_ATTR_READ  0x01		/* From memory */
207 #define PCI_MAP_ATTR_WRITE 0x02		/* To memory */
208 
209 /*
210  * PCI MSI services
211  */
212 
213 #ifndef _LOCORE
214 int64_t hv_pci_msiq_conf(uint64_t devhandle, uint64_t msiqid,
215 	    uint64_t r_addr, uint64_t nentries);
216 int64_t hv_pci_msiq_info(uint64_t devhandle, uint64_t msiqid,
217 	    uint64_t *r_addr, uint64_t *nentries);
218 
219 int64_t hv_pci_msiq_getvalid(uint64_t devhandle, uint64_t msiqid,
220 	    uint64_t *msiqvalid);
221 int64_t hv_pci_msiq_setvalid(uint64_t devhandle, uint64_t msiqid,
222 	    uint64_t msiqvalid);
223 #endif
224 
225 #define PCI_MSIQ_INVALID	0
226 #define PCI_MSIQ_VALID		1
227 
228 #ifndef _LOCORE
229 int64_t hv_pci_msiq_getstate(uint64_t devhandle, uint64_t msiqid,
230 	    uint64_t *msiqstate);
231 int64_t hv_pci_msiq_setstate(uint64_t devhandle, uint64_t msiqid,
232 	    uint64_t msiqstate);
233 #endif
234 
235 #define PCI_MSIQSTATE_IDLE	0
236 #define PCI_MSIQSTATE_ERROR	1
237 
238 #ifndef _LOCORE
239 int64_t hv_pci_msiq_gethead(uint64_t devhandle, uint64_t msiqid,
240 	    uint64_t *msiqhead);
241 int64_t hv_pci_msiq_sethead(uint64_t devhandle, uint64_t msiqid,
242 	    uint64_t msiqhead);
243 int64_t hv_pci_msiq_gettail(uint64_t devhandle, uint64_t msiqid,
244 	    uint64_t *msiqtail);
245 
246 int64_t hv_pci_msi_getvalid(uint64_t devhandle, uint64_t msinum,
247 	    uint64_t *msivalidstate);
248 int64_t hv_pci_msi_setvalid(uint64_t devhandle, uint64_t msinum,
249 	    uint64_t msivalidstate);
250 #endif
251 
252 #define PCI_MSI_INVALID		0
253 #define PCI_MSI_VALID		1
254 
255 #ifndef _LOCORE
256 int64_t hv_pci_msi_getmsiq(uint64_t devhandle, uint64_t msinum,
257 	    uint64_t *msiqid);
258 int64_t hv_pci_msi_setmsiq(uint64_t devhandle, uint64_t msinum,
259 	    uint64_t msitype, uint64_t msiqid);
260 
261 int64_t hv_pci_msi_getstate(uint64_t devhandle, uint64_t msinum,
262 	    uint64_t *msistate);
263 int64_t hv_pci_msi_setstate(uint64_t devhandle, uint64_t msinum,
264 	    uint64_t msistate);
265 #endif
266 
267 #define PCI_MSISTATE_IDLE	0
268 #define PCI_MSISTATE_DELIVERED	1
269 
270 #ifndef _LOCORE
271 int64_t hv_pci_msg_getmsiq(uint64_t devhandle, uint64_t msg,
272 	    uint64_t *msiqid);
273 int64_t hv_pci_msg_setmsiq(uint64_t devhandle, uint64_t msg,
274 	    uint64_t msiqid);
275 
276 int64_t hv_pci_msg_getvalid(uint64_t devhandle, uint64_t msg,
277 	    uint64_t *msgvalidstate);
278 int64_t hv_pci_msg_setvalid(uint64_t devhandle, uint64_t msg,
279 	    uint64_t msgvalidstate);
280 #endif
281 
282 #define PCIE_MSG_INVALID	0
283 #define PCIE_MSG_VALID		1
284 
285 #define PCIE_PME_MSG		0x18
286 #define PCIE_PME_ACK_MSG	0x1b
287 #define PCIE_CORR_MSG		0x30
288 #define PCIE_NONFATAL_MSG	0x31
289 #define PCIE_FATAL_MSG		0x32
290 
291 /*
292  * Logical Domain Channel services
293  */
294 
295 #ifndef _LOCORE
296 int64_t hv_ldc_tx_qconf(uint64_t ldc_id, paddr_t base_raddr,
297 	    uint64_t nentries);
298 int64_t hv_ldc_tx_qinfo(uint64_t ldc_id, paddr_t *base_raddr,
299 	    uint64_t *nentries);
300 int64_t hv_ldc_tx_get_state(uint64_t ldc_id, uint64_t *head_offset,
301 	    uint64_t *tail_offset, uint64_t *channel_state);
302 int64_t hv_ldc_tx_set_qtail(uint64_t ldc_id, uint64_t tail_offset);
303 int64_t hv_ldc_rx_qconf(uint64_t ldc_id, paddr_t base_raddr,
304 	    uint64_t nentries);
305 int64_t hv_ldc_rx_qinfo(uint64_t ldc_id, paddr_t *base_raddr,
306 	    uint64_t *nentries);
307 int64_t hv_ldc_rx_get_state(uint64_t ldc_id, uint64_t *head_offset,
308 	    uint64_t *tail_offset, uint64_t *channel_state);
309 int64_t hv_ldc_rx_set_qhead(uint64_t ldc_id, uint64_t head_offset);
310 #endif
311 
312 #define LDC_CHANNEL_DOWN	0
313 #define LDC_CHANNEL_UP		1
314 #define LDC_CHANNEL_RESET	2
315 
316 #ifndef _LOCORE
317 int64_t	hv_ldc_set_map_table(uint64_t ldc_id, paddr_t base_raddr,
318 	    uint64_t nentries);
319 int64_t	hv_ldc_get_map_table(uint64_t ldc_id, paddr_t *base_raddr,
320 	    uint64_t *nentries);
321 int64_t hv_ldc_copy(uint64_t ldc_id, uint64_t flags, uint64_t cookie,
322 	    paddr_t raddr, psize_t length, psize_t *ret_length);
323 #endif
324 
325 #define LDC_COPY_IN		0
326 #define LDC_COPY_OUT		1
327 
328 #ifndef _LOCORE
329 int64_t hv_ldc_mapin(uint64_t ldc_id, uint64_t cookie, paddr_t *raddr,
330 	    uint64_t *perms);
331 int64_t hv_ldc_unmap(paddr_t raddr, uint64_t *perms);
332 #endif
333 
334 /*
335  * Cryptographic services
336  */
337 
338 #ifndef _LOCORE
339 int64_t	hv_rng_get_diag_control(void);
340 int64_t	hv_rng_ctl_read(paddr_t raddr, uint64_t *state, uint64_t *delta);
341 int64_t	hv_rng_ctl_write(paddr_t raddr, uint64_t state, uint64_t timeout,
342 	uint64_t *delta);
343 #endif
344 
345 #define RNG_STATE_UNCONFIGURED	0
346 #define RNG_STATE_CONFIGURED	1
347 #define RNG_STATE_HEALTHCHECK	2
348 #define RNG_STATE_ERROR		3
349 
350 #ifndef _LOCORE
351 int64_t	hv_rng_data_read_diag(paddr_t raddr, uint64_t size, uint64_t *delta);
352 int64_t	hv_rng_data_read(paddr_t raddr, uint64_t *delta);
353 #endif
354 
355 /*
356  * Error codes
357  */
358 
359 #define H_EOK		0
360 #define H_ENOCPU	1
361 #define H_ENORADDR	2
362 #define H_ENOINTR	3
363 #define H_EBADPGSZ	4
364 #define H_EBADTSB	5
365 #define H_EINVAL	6
366 #define H_EBADTRAP	7
367 #define H_EBADALIGN	8
368 #define H_EWOULDBLOCK	9
369 #define H_ENOACCESS	10
370 #define H_EIO		11
371 #define H_ECPUERROR	12
372 #define H_ENOTSUPPORTED	13
373 #define H_ENOMAP	14
374 #define H_ETOOMANY	15
375 #define H_ECHANNEL	16
376 
377 #endif	/* _HYPERVISOR_H_ */
378