xref: /minix3/sys/arch/x86/include/cacheinfo.h (revision b1c4ba4ab66acbca4eb1bab1b49a15e96feef93e)
1 /*	$NetBSD: cacheinfo.h,v 1.13 2011/12/04 17:00:10 chs Exp $	*/
2 
3 #ifndef _X86_CACHEINFO_H_
4 #define _X86_CACHEINFO_H_
5 
6 struct x86_cache_info {
7 	uint8_t		cai_index;
8 	uint8_t		cai_desc;
9 	uint8_t		cai_associativity;
10 	u_int		cai_totalsize; /* #entries for TLB, bytes for cache */
11 	u_int		cai_linesize;	/* or page size for TLB */
12 #ifndef _KERNEL
13 	const char	*cai_string;
14 #endif
15 };
16 
17 #define	CAI_ITLB	0		/* Instruction TLB (4K pages) */
18 #define	CAI_ITLB2	1		/* Instruction TLB (2/4M pages) */
19 #define	CAI_DTLB	2		/* Data TLB (4K pages) */
20 #define	CAI_DTLB2	3		/* Data TLB (2/4M pages) */
21 #define	CAI_ICACHE	4		/* Instruction cache */
22 #define	CAI_DCACHE	5		/* Data cache */
23 #define	CAI_L2CACHE	6		/* Level 2 cache */
24 #define	CAI_L3CACHE	7		/* Level 3 cache */
25 #define	CAI_L1_1GBITLB	8		/* L1 1GB Page instruction TLB */
26 #define	CAI_L1_1GBDTLB	9		/* L1 1GB Page data TLB */
27 #define CAI_L2_1GBITLB	10		/* L2 1GB Page instruction TLB */
28 #define CAI_L2_1GBDTLB	11		/* L2 1GB Page data TLB */
29 #define CAI_L2_ITLB	12		/* L2 Instruction TLB (4K pages) */
30 #define CAI_L2_ITLB2	13		/* L2 Instruction TLB (2/4M pages) */
31 #define CAI_L2_DTLB	14		/* L2 Data TLB (4K pages) */
32 #define CAI_L2_DTLB2	15		/* L2 Data TLB (2/4M pages) */
33 
34 #define	CAI_COUNT	16
35 
36 /*
37  * AMD Cache Info:
38  *
39  *      Barcelona, Phenom:
40  *
41  *		Function 8000.0005 L1 TLB/Cache Information
42  *		EAX -- L1 TLB 2/4MB pages
43  *		EBX -- L1 TLB 4K pages
44  *		ECX -- L1 D-cache
45  *		EDX -- L1 I-cache
46  *
47  *		Function 8000.0006 L2 TLB/Cache Information
48  *		EAX -- L2 TLB 2/4MB pages
49  *		EBX -- L2 TLB 4K pages
50  *		ECX -- L2 Unified cache
51  *		EDX -- L3 Unified Cache
52  *
53  *		Function 8000.0019 TLB 1GB Page Information
54  *		EAX -- L1 1GB pages
55  *		EBX -- L2 1GB pages
56  *		ECX -- reserved
57  *		EDX -- reserved
58  *
59  *	Athlon, Duron:
60  *
61  *		Function 8000.0005 L1 TLB/Cache Information
62  *		EAX -- L1 TLB 2/4MB pages
63  *		EBX -- L1 TLB 4K pages
64  *		ECX -- L1 D-cache
65  *		EDX -- L1 I-cache
66  *
67  *		Function 8000.0006 L2 TLB/Cache Information
68  *		EAX -- L2 TLB 2/4MB pages
69  *		EBX -- L2 TLB 4K pages
70  *		ECX -- L2 Unified cache
71  *		EDX -- reserved
72  *
73  *	K5, K6:
74  *
75  *		Function 8000.0005 L1 TLB/Cache Information
76  *		EAX -- reserved
77  *		EBX -- TLB 4K pages
78  *		ECX -- L1 D-cache
79  *		EDX -- L1 I-cache
80  *
81  *	K6-III:
82  *
83  *		Function 8000.0006 L2 Cache Information
84  *		EAX -- reserved
85  *		EBX -- reserved
86  *		ECX -- L2 Unified cache
87  *		EDX -- reserved
88  */
89 
90 /* L1 TLB 2/4MB pages */
91 #define	AMD_L1_EAX_DTLB_ASSOC(x)	(((x) >> 24) & 0xff)
92 #define	AMD_L1_EAX_DTLB_ENTRIES(x)	(((x) >> 16) & 0xff)
93 #define	AMD_L1_EAX_ITLB_ASSOC(x)	(((x) >> 8)  & 0xff)
94 #define	AMD_L1_EAX_ITLB_ENTRIES(x)	( (x)        & 0xff)
95 
96 /* L1 TLB 4K pages */
97 #define	AMD_L1_EBX_DTLB_ASSOC(x)	(((x) >> 24) & 0xff)
98 #define	AMD_L1_EBX_DTLB_ENTRIES(x)	(((x) >> 16) & 0xff)
99 #define	AMD_L1_EBX_ITLB_ASSOC(x)	(((x) >> 8)  & 0xff)
100 #define	AMD_L1_EBX_ITLB_ENTRIES(x)	( (x)        & 0xff)
101 
102 /* L1 Data Cache */
103 #define	AMD_L1_ECX_DC_SIZE(x)		((((x) >> 24) & 0xff) * 1024)
104 #define	AMD_L1_ECX_DC_ASSOC(x)		 (((x) >> 16) & 0xff)
105 #define	AMD_L1_ECX_DC_LPT(x)		 (((x) >> 8)  & 0xff)
106 #define	AMD_L1_ECX_DC_LS(x)		 ( (x)        & 0xff)
107 
108 /* L1 Instruction Cache */
109 #define	AMD_L1_EDX_IC_SIZE(x)		((((x) >> 24) & 0xff) * 1024)
110 #define	AMD_L1_EDX_IC_ASSOC(x)		 (((x) >> 16) & 0xff)
111 #define	AMD_L1_EDX_IC_LPT(x)		 (((x) >> 8)  & 0xff)
112 #define	AMD_L1_EDX_IC_LS(x)		 ( (x)        & 0xff)
113 
114 /* Note for L2 TLB -- if the upper 16 bits are 0, it is a unified TLB */
115 
116 /* L2 TLB 2/4MB pages */
117 #define	AMD_L2_EAX_DTLB_ASSOC(x)	(((x) >> 28)  & 0xf)
118 #define	AMD_L2_EAX_DTLB_ENTRIES(x)	(((x) >> 16)  & 0xfff)
119 #define	AMD_L2_EAX_IUTLB_ASSOC(x)	(((x) >> 12)  & 0xf)
120 #define	AMD_L2_EAX_IUTLB_ENTRIES(x)	( (x)         & 0xfff)
121 
122 /* L2 TLB 4K pages */
123 #define	AMD_L2_EBX_DTLB_ASSOC(x)	(((x) >> 28)  & 0xf)
124 #define	AMD_L2_EBX_DTLB_ENTRIES(x)	(((x) >> 16)  & 0xfff)
125 #define	AMD_L2_EBX_IUTLB_ASSOC(x)	(((x) >> 12)  & 0xf)
126 #define	AMD_L2_EBX_IUTLB_ENTRIES(x)	( (x)         & 0xfff)
127 
128 /* L2 Cache */
129 #define	AMD_L2_ECX_C_SIZE(x)		((((x) >> 16) & 0xffff) * 1024)
130 #define	AMD_L2_ECX_C_ASSOC(x)		 (((x) >> 12) & 0xf)
131 #define	AMD_L2_ECX_C_LPT(x)		 (((x) >> 8)  & 0xf)
132 #define	AMD_L2_ECX_C_LS(x)		 ( (x)        & 0xff)
133 
134 /* L3 Cache */
135 #define AMD_L3_EDX_C_SIZE(x)		((((x) >> 18) & 0xffff) * 1024 * 512)
136 #define AMD_L3_EDX_C_ASSOC(x)		 (((x) >> 12) & 0xff)
137 #define AMD_L3_EDX_C_LPT(x)		 (((x) >> 8)  & 0xf)
138 #define AMD_L3_EDX_C_LS(x)		 ( (x)        & 0xff)
139 
140 /* L1 TLB 1GB pages */
141 #define AMD_L1_1GB_EAX_DTLB_ASSOC(x)	(((x) >> 28) & 0xf)
142 #define AMD_L1_1GB_EAX_DTLB_ENTRIES(x)	(((x) >> 16) & 0xfff)
143 #define AMD_L1_1GB_EAX_IUTLB_ASSOC(x)	(((x) >> 12) & 0xf)
144 #define AMD_L1_1GB_EAX_IUTLB_ENTRIES(x)	( (x)        & 0xfff)
145 
146 /* L2 TLB 1GB pages */
147 #define AMD_L2_1GB_EBX_DUTLB_ASSOC(x)	(((x) >> 28) & 0xf)
148 #define AMD_L2_1GB_EBX_DUTLB_ENTRIES(x)	(((x) >> 16) & 0xfff)
149 #define AMD_L2_1GB_EBX_IUTLB_ASSOC(x)	(((x) >> 12) & 0xf)
150 #define AMD_L2_1GB_EBX_IUTLB_ENTRIES(x)	( (x)        & 0xfff)
151 
152 /*
153  * VIA Cache Info:
154  *
155  *	Nehemiah (at least)
156  *
157  *		Function 8000.0005 L1 TLB/Cache Information
158  *		EAX -- reserved
159  *		EBX -- L1 TLB 4K pages
160  *		ECX -- L1 D-cache
161  *		EDX -- L1 I-cache
162  *
163  *		Function 8000.0006 L2 Cache Information
164  *		EAX -- reserved
165  *		EBX -- reserved
166  *		ECX -- L2 Unified cache
167  *		EDX -- reserved
168  */
169 
170 /* L1 TLB 4K pages */
171 #define	VIA_L1_EBX_DTLB_ASSOC(x)	(((x) >> 24) & 0xff)
172 #define	VIA_L1_EBX_DTLB_ENTRIES(x)	(((x) >> 16) & 0xff)
173 #define	VIA_L1_EBX_ITLB_ASSOC(x)	(((x) >> 8)  & 0xff)
174 #define	VIA_L1_EBX_ITLB_ENTRIES(x)	( (x)        & 0xff)
175 
176 /* L1 Data Cache */
177 #define	VIA_L1_ECX_DC_SIZE(x)		((((x) >> 24) & 0xff) * 1024)
178 #define	VIA_L1_ECX_DC_ASSOC(x)		 (((x) >> 16) & 0xff)
179 #define	VIA_L1_ECX_DC_LPT(x)		 (((x) >> 8)  & 0xff)
180 #define	VIA_L1_ECX_DC_LS(x)		 ( (x)        & 0xff)
181 
182 /* L1 Instruction Cache */
183 #define	VIA_L1_EDX_IC_SIZE(x)		((((x) >> 24) & 0xff) * 1024)
184 #define	VIA_L1_EDX_IC_ASSOC(x)		 (((x) >> 16) & 0xff)
185 #define	VIA_L1_EDX_IC_LPT(x)		 (((x) >> 8)  & 0xff)
186 #define	VIA_L1_EDX_IC_LS(x)		 ( (x)        & 0xff)
187 
188 /* L2 Cache (pre-Nehemiah) */
189 #define	VIA_L2_ECX_C_SIZE(x)		((((x) >> 24) & 0xff) * 1024)
190 #define	VIA_L2_ECX_C_ASSOC(x)		 (((x) >> 16) & 0xff)
191 #define	VIA_L2_ECX_C_LPT(x)		 (((x) >> 8)  & 0xff)
192 #define	VIA_L2_ECX_C_LS(x)		 ( (x)        & 0xff)
193 
194 /* L2 Cache (Nehemiah and newer) */
195 #define	VIA_L2N_ECX_C_SIZE(x)		((((x) >> 16) & 0xffff) * 1024)
196 #define	VIA_L2N_ECX_C_ASSOC(x)		 (((x) >> 12) & 0xf)
197 #define	VIA_L2N_ECX_C_LPT(x)		 (((x) >> 8)  & 0xf)
198 #define	VIA_L2N_ECX_C_LS(x)		 ( (x)        & 0xff)
199 
200 #ifdef _KERNEL
201 #define __CI_TBL(a,b,c,d,e,f) { a, b, c, d, e }
202 #else
203 #define __CI_TBL(a,b,c,d,e,f) { a, b, c, d, e, f }
204 #endif
205 
206 /*
207  * XXX Currently organized mostly by cache type, but would be
208  * XXX easier to maintain if it were in descriptor type order.
209  */
210 #define INTEL_CACHE_INFO { \
211 __CI_TBL(CAI_ITLB,     0x01,    4, 32,        4 * 1024, NULL), \
212 __CI_TBL(CAI_ITLB,     0xb0,    4,128,        4 * 1024, NULL), \
213 __CI_TBL(CAI_ITLB2,    0x02, 0xff,  2, 4 * 1024 * 1024, NULL), \
214 __CI_TBL(CAI_DTLB,     0x03,    4, 64,        4 * 1024, NULL), \
215 __CI_TBL(CAI_DTLB,     0xb3,    4,128,        4 * 1024, NULL), \
216 __CI_TBL(CAI_DTLB,     0xb4,    4,256,        4 * 1024, NULL), \
217 __CI_TBL(CAI_DTLB2,    0x04,    4,  8, 4 * 1024 * 1024, NULL), \
218 __CI_TBL(CAI_DTLB2,    0x05,    4, 32, 4 * 1024 * 1024, NULL), \
219 __CI_TBL(CAI_ITLB,     0x50, 0xff, 64,        4 * 1024, "4K/4M: 64 entries"), \
220 __CI_TBL(CAI_ITLB,     0x51, 0xff, 64,        4 * 1024, "4K/4M: 128 entries"),\
221 __CI_TBL(CAI_ITLB,     0x52, 0xff, 64,        4 * 1024, "4K/4M: 256 entries"),\
222 __CI_TBL(CAI_ITLB,     0x55, 0xff, 64,        4 * 1024, "2M/4M: 7 entries"), \
223 __CI_TBL(CAI_DTLB2,    0x56,    4, 16, 4 * 1024 * 1024, NULL), \
224 __CI_TBL(CAI_DTLB2,    0x57,    4, 16,        4 * 1024, NULL), \
225 __CI_TBL(CAI_DTLB,     0x5a, 0xff, 64,        4 * 1024, "2M/4M: 32 entries (L0)"), \
226 __CI_TBL(CAI_DTLB,     0x5b, 0xff, 64,        4 * 1024, "4K/4M: 64 entries"), \
227 __CI_TBL(CAI_DTLB,     0x5c, 0xff, 64,        4 * 1024, "4K/4M: 128 entries"),\
228 __CI_TBL(CAI_DTLB,     0x5d, 0xff, 64,        4 * 1024, "4K/4M: 256 entries"),\
229 __CI_TBL(CAI_ITLB,     0xb1,    4, 64,               0, "8 2M/4 4M entries"), \
230 __CI_TBL(CAI_ITLB,     0xb2,    4, 64,        4 * 1024, NULL), \
231 __CI_TBL(CAI_ICACHE,   0x06,    4,        8 * 1024, 32, NULL), \
232 __CI_TBL(CAI_ICACHE,   0x08,    4,       16 * 1024, 32, NULL), \
233 __CI_TBL(CAI_ICACHE,   0x09,    4,       32 * 1024, 64, NULL), \
234 __CI_TBL(CAI_ICACHE,   0x30,    8,       32 * 1024, 64, NULL), \
235 __CI_TBL(CAI_DCACHE,   0x0a,    2,        8 * 1024, 32, NULL), \
236 __CI_TBL(CAI_DCACHE,   0x0c,    4,       16 * 1024, 32, NULL), \
237 __CI_TBL(CAI_DCACHE,   0x0d,    4,       16 * 1024, 32, NULL), \
238 __CI_TBL(CAI_L2CACHE,  0x21,    8,      256 * 1024, 64, NULL), /* L2 (MLC) */ \
239 __CI_TBL(CAI_L2CACHE,  0x39,    4,      128 * 1024, 64, NULL), \
240 __CI_TBL(CAI_L2CACHE,  0x3a,    6,      192 * 1024, 64, NULL), \
241 __CI_TBL(CAI_L2CACHE,  0x3b,    2,      128 * 1024, 64, NULL), \
242 __CI_TBL(CAI_L2CACHE,  0x3c,    4,      256 * 1024, 64, NULL), \
243 __CI_TBL(CAI_L2CACHE,  0x3d,    6,      384 * 1024, 64, NULL), \
244 __CI_TBL(CAI_L2CACHE,  0x3e,    4,      512 * 1024, 64, NULL), \
245 __CI_TBL(CAI_L2CACHE,  0x40,    0,               0,  0, "not present"), \
246 __CI_TBL(CAI_L2CACHE,  0x41,    4,      128 * 1024, 32, NULL), \
247 __CI_TBL(CAI_L2CACHE,  0x42,    4,      256 * 1024, 32, NULL), \
248 __CI_TBL(CAI_L2CACHE,  0x43,    4,      512 * 1024, 32, NULL), \
249 __CI_TBL(CAI_L2CACHE,  0x44,    4, 1 * 1024 * 1024, 32, NULL), \
250 __CI_TBL(CAI_L2CACHE,  0x45,    4, 2 * 1024 * 1024, 32, NULL), \
251 __CI_TBL(CAI_L2CACHE,  0x48,   12, 3 * 1024 * 1024, 64, NULL), \
252 								\
253 /* 0x49 Is L2 on Xeon MP (Family 0f, Model 06), L3 otherwise */	\
254 __CI_TBL(CAI_L2CACHE,  0x49,   16, 4 * 1024 * 1024, 64, NULL), \
255 __CI_TBL(CAI_L2CACHE,  0x4e,   24, 6 * 1024 * 1024, 64, NULL), \
256 __CI_TBL(CAI_DCACHE,   0x60,    8,       16 * 1024, 64, NULL), \
257 __CI_TBL(CAI_DCACHE,   0x66,    4,        8 * 1024, 64, NULL), \
258 __CI_TBL(CAI_DCACHE,   0x67,    4,       16 * 1024, 64, NULL), \
259 __CI_TBL(CAI_DCACHE,   0x2c,    8,       32 * 1024, 64, NULL), \
260 __CI_TBL(CAI_DCACHE,   0x68,    4,       32 * 1024, 64, NULL), \
261 __CI_TBL(CAI_ICACHE,   0x70,    8,       12 * 1024, 64, "12K uOp cache"), \
262 __CI_TBL(CAI_ICACHE,   0x71,    8,       16 * 1024, 64, "16K uOp cache"), \
263 __CI_TBL(CAI_ICACHE,   0x72,    8,       32 * 1024, 64, "32K uOp cache"), \
264 __CI_TBL(CAI_ICACHE,   0x73,    8,       64 * 1024, 64, "64K uOp cache"), \
265 __CI_TBL(CAI_L2CACHE,  0x78,    4, 1 * 1024 * 1024, 64, NULL), \
266 __CI_TBL(CAI_L2CACHE,  0x79,    8,      128 * 1024, 64, NULL), \
267 __CI_TBL(CAI_L2CACHE,  0x7a,    8,      256 * 1024, 64, NULL), \
268 __CI_TBL(CAI_L2CACHE,  0x7b,    8,      512 * 1024, 64, NULL), \
269 __CI_TBL(CAI_L2CACHE,  0x7c,    8, 1 * 1024 * 1024, 64, NULL), \
270 __CI_TBL(CAI_L2CACHE,  0x7d,    8, 2 * 1024 * 1024, 64, NULL), \
271 __CI_TBL(CAI_L2CACHE,  0x7f,    2,      512 * 1024, 64, NULL), \
272 __CI_TBL(CAI_L2CACHE,  0x82,    8,      256 * 1024, 32, NULL), \
273 __CI_TBL(CAI_L2CACHE,  0x83,    8,      512 * 1024, 32, NULL), \
274 __CI_TBL(CAI_L2CACHE,  0x84,    8, 1 * 1024 * 1024, 32, NULL), \
275 __CI_TBL(CAI_L2CACHE,  0x85,    8, 2 * 1024 * 1024, 32, NULL), \
276 __CI_TBL(CAI_L2CACHE,  0x86,    4,      512 * 1024, 64, NULL), \
277 __CI_TBL(CAI_L2CACHE,  0x87,    8, 1 * 1024 * 1024, 64, NULL), \
278 __CI_TBL(CAI_L3CACHE,  0x22, 0xff,      512 * 1024, 64, "sectored, 4-way "), \
279 __CI_TBL(CAI_L3CACHE,  0x23, 0xff, 1 * 1024 * 1024, 64, "sectored, 8-way "), \
280 __CI_TBL(CAI_L3CACHE,  0x25, 0xff, 2 * 1024 * 1024, 64, "sectored, 8-way "), \
281 __CI_TBL(CAI_L3CACHE,  0x29, 0xff, 4 * 1024 * 1024, 64, "sectored, 8-way "), \
282 __CI_TBL(CAI_L3CACHE,  0x46,    4, 4 * 1024 * 1024, 64, NULL), \
283 __CI_TBL(CAI_L3CACHE,  0x47,    8, 8 * 1024 * 1024, 64, NULL), \
284 __CI_TBL(CAI_L3CACHE,  0x49,   16, 4 * 1024 * 1024, 64, NULL), \
285 __CI_TBL(CAI_L3CACHE,  0x4a,   12, 6 * 1024 * 1024, 64, NULL), \
286 __CI_TBL(CAI_L3CACHE,  0x4b,   16, 8 * 1024 * 1024, 64, NULL), \
287 __CI_TBL(CAI_L3CACHE,  0x4c,   12,12 * 1024 * 1024, 64, NULL), \
288 __CI_TBL(CAI_L3CACHE,  0x4d,   16,16 * 1024 * 1024, 64, NULL), \
289 __CI_TBL(CAI_L3CACHE,  0xd0,    4,      512 * 1024, 64, NULL), \
290 __CI_TBL(CAI_L3CACHE,  0xd1,    4, 1 * 1024 * 1024, 64, NULL), \
291 __CI_TBL(CAI_L3CACHE,  0xd2,    4, 2 * 1024 * 1024, 64, NULL), \
292 __CI_TBL(CAI_L3CACHE,  0xd6,    8, 1 * 1024 * 1024, 64, NULL), \
293 __CI_TBL(CAI_L3CACHE,  0xd7,    8, 2 * 1024 * 1024, 64, NULL), \
294 __CI_TBL(CAI_L3CACHE,  0xd8,    8, 4 * 1024 * 1024, 64, NULL), \
295 __CI_TBL(CAI_L3CACHE,  0xdc,   12, 3 *  512 * 1024, 64, NULL), \
296 __CI_TBL(CAI_L3CACHE,  0xdd,   12, 3 * 1024 * 1024, 64, NULL), \
297 __CI_TBL(CAI_L3CACHE,  0xde,   12, 6 * 1024 * 1024, 64, NULL), \
298 __CI_TBL(CAI_L3CACHE,  0xe2,   16, 2 * 1024 * 1024, 64, NULL), \
299 __CI_TBL(CAI_L3CACHE,  0xe3,   16, 4 * 1024 * 1024, 64, NULL), \
300 __CI_TBL(CAI_L3CACHE,  0xe4,   16, 8 * 1024 * 1024, 64, NULL), \
301 __CI_TBL(CAI_L3CACHE,  0xea,   24,12 * 1024 * 1024, 64, NULL), \
302 __CI_TBL(CAI_L3CACHE,  0xeb,   24,24 * 1024 * 1024, 64, NULL), \
303 __CI_TBL(CAI_L3CACHE,  0xec,   24,24 * 1024 * 1024, 64, NULL), \
304 __CI_TBL(0,               0,    0,               0,  0, NULL)  \
305 }
306 
307 #define AMD_L2CACHE_INFO { \
308 __CI_TBL(0, 0x01,    1, 0, 0, NULL), \
309 __CI_TBL(0, 0x02,    2, 0, 0, NULL), \
310 __CI_TBL(0, 0x04,    4, 0, 0, NULL), \
311 __CI_TBL(0, 0x06,    8, 0, 0, NULL), \
312 __CI_TBL(0, 0x08,   16, 0, 0, NULL), \
313 __CI_TBL(0, 0x0a,   32, 0, 0, NULL), \
314 __CI_TBL(0, 0x0b,   48, 0, 0, NULL), \
315 __CI_TBL(0, 0x0c,   64, 0, 0, NULL), \
316 __CI_TBL(0, 0x0d,   96, 0, 0, NULL), \
317 __CI_TBL(0, 0x0e,  128, 0, 0, NULL), \
318 __CI_TBL(0, 0x0f, 0xff, 0, 0, NULL), \
319 __CI_TBL(0, 0x00,    0, 0, 0, NULL)  \
320 }
321 
322 #define AMD_L3CACHE_INFO { \
323 __CI_TBL(0, 0x01,    1, 0, 0, NULL), \
324 __CI_TBL(0, 0x02,    2, 0, 0, NULL), \
325 __CI_TBL(0, 0x04,    4, 0, 0, NULL), \
326 __CI_TBL(0, 0x06,    8, 0, 0, NULL), \
327 __CI_TBL(0, 0x08,   16, 0, 0, NULL), \
328 __CI_TBL(0, 0x0a,   32, 0, 0, NULL), \
329 __CI_TBL(0, 0x0b,   48, 0, 0, NULL), \
330 __CI_TBL(0, 0x0c,   64, 0, 0, NULL), \
331 __CI_TBL(0, 0x0d,   96, 0, 0, NULL), \
332 __CI_TBL(0, 0x0e,  128, 0, 0, NULL), \
333 __CI_TBL(0, 0x0f, 0xff, 0, 0, NULL), \
334 __CI_TBL(0, 0x00,    0, 0, 0, NULL)  \
335 }
336 
337 #endif /* _X86_CACHEINFO_H_ */
338