Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / arch / ia64 / include / asm / paravirt_privop.h
1 /******************************************************************************
2  * Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
3  *                    VA Linux Systems Japan K.K.
4  *
5  * This program is free software; you can redistribute it and/or modify
6  * it under the terms of the GNU General Public License as published by
7  * the Free Software Foundation; either version 2 of the License, or
8  * (at your option) any later version.
9  *
10  * This program is distributed in the hope that it will be useful,
11  * but WITHOUT ANY WARRANTY; without even the implied warranty of
12  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
13  * GNU General Public License for more details.
14  *
15  * You should have received a copy of the GNU General Public License
16  * along with this program; if not, write to the Free Software
17  * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
18  *
19  */
20
21 #ifndef _ASM_IA64_PARAVIRT_PRIVOP_H
22 #define _ASM_IA64_PARAVIRT_PRIVOP_H
23
24 #ifdef CONFIG_PARAVIRT
25
26 #ifndef __ASSEMBLY__
27
28 #include <linux/types.h>
29 #include <asm/kregs.h> /* for IA64_PSR_I */
30
31 /******************************************************************************
32  * replacement of intrinsics operations.
33  */
34
35 struct pv_cpu_ops {
36         void (*fc)(void *addr);
37         unsigned long (*thash)(unsigned long addr);
38         unsigned long (*get_cpuid)(int index);
39         unsigned long (*get_pmd)(int index);
40         unsigned long (*getreg)(int reg);
41         void (*setreg)(int reg, unsigned long val);
42         void (*ptcga)(unsigned long addr, unsigned long size);
43         unsigned long (*get_rr)(unsigned long index);
44         void (*set_rr)(unsigned long index, unsigned long val);
45         void (*set_rr0_to_rr4)(unsigned long val0, unsigned long val1,
46                                unsigned long val2, unsigned long val3,
47                                unsigned long val4);
48         void (*ssm_i)(void);
49         void (*rsm_i)(void);
50         unsigned long (*get_psr_i)(void);
51         void (*intrin_local_irq_restore)(unsigned long flags);
52 };
53
54 extern struct pv_cpu_ops pv_cpu_ops;
55
56 extern void ia64_native_setreg_func(int regnum, unsigned long val);
57 extern unsigned long ia64_native_getreg_func(int regnum);
58
59 /************************************************/
60 /* Instructions paravirtualized for performance */
61 /************************************************/
62
63 #ifndef ASM_SUPPORTED
64 #define paravirt_ssm_i()        pv_cpu_ops.ssm_i()
65 #define paravirt_rsm_i()        pv_cpu_ops.rsm_i()
66 #define __paravirt_getreg()     pv_cpu_ops.getreg()
67 #endif
68
69 /* mask for ia64_native_ssm/rsm() must be constant.("i" constraing).
70  * static inline function doesn't satisfy it. */
71 #define paravirt_ssm(mask)                      \
72         do {                                    \
73                 if ((mask) == IA64_PSR_I)       \
74                         paravirt_ssm_i();       \
75                 else                            \
76                         ia64_native_ssm(mask);  \
77         } while (0)
78
79 #define paravirt_rsm(mask)                      \
80         do {                                    \
81                 if ((mask) == IA64_PSR_I)       \
82                         paravirt_rsm_i();       \
83                 else                            \
84                         ia64_native_rsm(mask);  \
85         } while (0)
86
87 /* returned ip value should be the one in the caller,
88  * not in __paravirt_getreg() */
89 #define paravirt_getreg(reg)                                    \
90         ({                                                      \
91                 unsigned long res;                              \
92                 if ((reg) == _IA64_REG_IP)                      \
93                         res = ia64_native_getreg(_IA64_REG_IP); \
94                 else                                            \
95                         res = __paravirt_getreg(reg);           \
96                 res;                                            \
97         })
98
99 /******************************************************************************
100  * replacement of hand written assembly codes.
101  */
102 struct pv_cpu_asm_switch {
103         unsigned long switch_to;
104         unsigned long leave_syscall;
105         unsigned long work_processed_syscall;
106         unsigned long leave_kernel;
107 };
108 void paravirt_cpu_asm_init(const struct pv_cpu_asm_switch *cpu_asm_switch);
109
110 #endif /* __ASSEMBLY__ */
111
112 #define IA64_PARAVIRT_ASM_FUNC(name)    paravirt_ ## name
113
114 #else
115
116 /* fallback for native case */
117 #define IA64_PARAVIRT_ASM_FUNC(name)    ia64_native_ ## name
118
119 #endif /* CONFIG_PARAVIRT */
120
121 #if defined(CONFIG_PARAVIRT) && defined(ASM_SUPPORTED)
122 #define paravirt_dv_serialize_data()    ia64_dv_serialize_data()
123 #else
124 #define paravirt_dv_serialize_data()    /* nothing */
125 #endif
126
127 /* these routines utilize privilege-sensitive or performance-sensitive
128  * privileged instructions so the code must be replaced with
129  * paravirtualized versions */
130 #define ia64_switch_to                  IA64_PARAVIRT_ASM_FUNC(switch_to)
131 #define ia64_leave_syscall              IA64_PARAVIRT_ASM_FUNC(leave_syscall)
132 #define ia64_work_processed_syscall     \
133         IA64_PARAVIRT_ASM_FUNC(work_processed_syscall)
134 #define ia64_leave_kernel               IA64_PARAVIRT_ASM_FUNC(leave_kernel)
135
136
137 #if defined(CONFIG_PARAVIRT)
138 /******************************************************************************
139  * binary patching infrastructure
140  */
141 #define PARAVIRT_PATCH_TYPE_FC                          1
142 #define PARAVIRT_PATCH_TYPE_THASH                       2
143 #define PARAVIRT_PATCH_TYPE_GET_CPUID                   3
144 #define PARAVIRT_PATCH_TYPE_GET_PMD                     4
145 #define PARAVIRT_PATCH_TYPE_PTCGA                       5
146 #define PARAVIRT_PATCH_TYPE_GET_RR                      6
147 #define PARAVIRT_PATCH_TYPE_SET_RR                      7
148 #define PARAVIRT_PATCH_TYPE_SET_RR0_TO_RR4              8
149 #define PARAVIRT_PATCH_TYPE_SSM_I                       9
150 #define PARAVIRT_PATCH_TYPE_RSM_I                       10
151 #define PARAVIRT_PATCH_TYPE_GET_PSR_I                   11
152 #define PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE    12
153
154 /* PARAVIRT_PATY_TYPE_[GS]ETREG + _IA64_REG_xxx */
155 #define PARAVIRT_PATCH_TYPE_GETREG                      0x10000000
156 #define PARAVIRT_PATCH_TYPE_SETREG                      0x20000000
157
158 /*
159  * struct task_struct* (*ia64_switch_to)(void* next_task);
160  * void *ia64_leave_syscall;
161  * void *ia64_work_processed_syscall
162  * void *ia64_leave_kernel;
163  */
164
165 #define PARAVIRT_PATCH_TYPE_BR_START                    0x30000000
166 #define PARAVIRT_PATCH_TYPE_BR_SWITCH_TO                \
167         (PARAVIRT_PATCH_TYPE_BR_START + 0)
168 #define PARAVIRT_PATCH_TYPE_BR_LEAVE_SYSCALL            \
169         (PARAVIRT_PATCH_TYPE_BR_START + 1)
170 #define PARAVIRT_PATCH_TYPE_BR_WORK_PROCESSED_SYSCALL   \
171         (PARAVIRT_PATCH_TYPE_BR_START + 2)
172 #define PARAVIRT_PATCH_TYPE_BR_LEAVE_KERNEL             \
173         (PARAVIRT_PATCH_TYPE_BR_START + 3)
174
175 #ifdef ASM_SUPPORTED
176 #include <asm/paravirt_patch.h>
177
178 /*
179  * pv_cpu_ops calling stub.
180  * normal function call convension can't be written by gcc
181  * inline assembly.
182  *
183  * from the caller's point of view,
184  * the following registers will be clobbered.
185  * r2, r3
186  * r8-r15
187  * r16, r17
188  * b6, b7
189  * p6-p15
190  * ar.ccv
191  *
192  * from the callee's point of view ,
193  * the following registers can be used.
194  * r2, r3: scratch
195  * r8: scratch, input argument0 and return value
196  * r0-r15: scratch, input argument1-5
197  * b6: return pointer
198  * b7: scratch
199  * p6-p15: scratch
200  * ar.ccv: scratch
201  *
202  * other registers must not be changed. especially
203  * b0: rp: preserved. gcc ignores b0 in clobbered register.
204  * r16: saved gp
205  */
206 /* 5 bundles */
207 #define __PARAVIRT_BR                                                   \
208         ";;\n"                                                          \
209         "{ .mlx\n"                                                      \
210         "nop 0\n"                                                       \
211         "movl r2 = %[op_addr]\n"/* get function pointer address */      \
212         ";;\n"                                                          \
213         "}\n"                                                           \
214         "1:\n"                                                          \
215         "{ .mii\n"                                                      \
216         "ld8 r2 = [r2]\n"       /* load function descriptor address */  \
217         "mov r17 = ip\n"        /* get ip to calc return address */     \
218         "mov r16 = gp\n"        /* save gp */                           \
219         ";;\n"                                                          \
220         "}\n"                                                           \
221         "{ .mii\n"                                                      \
222         "ld8 r3 = [r2], 8\n"    /* load entry address */                \
223         "adds r17 =  1f - 1b, r17\n"    /* calculate return address */  \
224         ";;\n"                                                          \
225         "mov b7 = r3\n"         /* set entry address */                 \
226         "}\n"                                                           \
227         "{ .mib\n"                                                      \
228         "ld8 gp = [r2]\n"       /* load gp value */                     \
229         "mov b6 = r17\n"        /* set return address */                \
230         "br.cond.sptk.few b7\n" /* intrinsics are very short isns */    \
231         "}\n"                                                           \
232         "1:\n"                                                          \
233         "{ .mii\n"                                                      \
234         "mov gp = r16\n"        /* restore gp value */                  \
235         "nop 0\n"                                                       \
236         "nop 0\n"                                                       \
237         ";;\n"                                                          \
238         "}\n"
239
240 #define PARAVIRT_OP(op)                         \
241         [op_addr] "i"(&pv_cpu_ops.op)
242
243 #define PARAVIRT_TYPE(type)                     \
244         PARAVIRT_PATCH_TYPE_ ## type
245
246 #define PARAVIRT_REG_CLOBBERS0                                  \
247         "r2", "r3", /*"r8",*/ "r9", "r10", "r11", "r14",        \
248                 "r15", "r16", "r17"
249
250 #define PARAVIRT_REG_CLOBBERS1                                  \
251         "r2","r3", /*"r8",*/ "r9", "r10", "r11", "r14",         \
252                 "r15", "r16", "r17"
253
254 #define PARAVIRT_REG_CLOBBERS2                                  \
255         "r2", "r3", /*"r8", "r9",*/ "r10", "r11", "r14",        \
256                 "r15", "r16", "r17"
257
258 #define PARAVIRT_REG_CLOBBERS5                                  \
259         "r2", "r3", /*"r8", "r9", "r10", "r11", "r14",*/        \
260                 "r15", "r16", "r17"
261
262 #define PARAVIRT_BR_CLOBBERS                    \
263         "b6", "b7"
264
265 #define PARAVIRT_PR_CLOBBERS                                            \
266         "p6", "p7", "p8", "p9", "p10", "p11", "p12", "p13", "p14", "p15"
267
268 #define PARAVIRT_AR_CLOBBERS                    \
269         "ar.ccv"
270
271 #define PARAVIRT_CLOBBERS0                      \
272                 PARAVIRT_REG_CLOBBERS0,         \
273                 PARAVIRT_BR_CLOBBERS,           \
274                 PARAVIRT_PR_CLOBBERS,           \
275                 PARAVIRT_AR_CLOBBERS,           \
276                 "memory"
277
278 #define PARAVIRT_CLOBBERS1                      \
279                 PARAVIRT_REG_CLOBBERS1,         \
280                 PARAVIRT_BR_CLOBBERS,           \
281                 PARAVIRT_PR_CLOBBERS,           \
282                 PARAVIRT_AR_CLOBBERS,           \
283                 "memory"
284
285 #define PARAVIRT_CLOBBERS2                      \
286                 PARAVIRT_REG_CLOBBERS2,         \
287                 PARAVIRT_BR_CLOBBERS,           \
288                 PARAVIRT_PR_CLOBBERS,           \
289                 PARAVIRT_AR_CLOBBERS,           \
290                 "memory"
291
292 #define PARAVIRT_CLOBBERS5                      \
293                 PARAVIRT_REG_CLOBBERS5,         \
294                 PARAVIRT_BR_CLOBBERS,           \
295                 PARAVIRT_PR_CLOBBERS,           \
296                 PARAVIRT_AR_CLOBBERS,           \
297                 "memory"
298
299 #define PARAVIRT_BR0(op, type)                                  \
300         register unsigned long ia64_clobber asm ("r8");         \
301         asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,        \
302                                           PARAVIRT_TYPE(type))  \
303                       : "=r"(ia64_clobber)                      \
304                       : PARAVIRT_OP(op)                         \
305                       : PARAVIRT_CLOBBERS0)
306
307 #define PARAVIRT_BR0_RET(op, type)                              \
308         register unsigned long ia64_intri_res asm ("r8");       \
309         asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,        \
310                                           PARAVIRT_TYPE(type))  \
311                       : "=r"(ia64_intri_res)                    \
312                       : PARAVIRT_OP(op)                         \
313                       : PARAVIRT_CLOBBERS0)
314
315 #define PARAVIRT_BR1(op, type, arg1)                            \
316         register unsigned long __##arg1 asm ("r8") = arg1;      \
317         register unsigned long ia64_clobber asm ("r8");         \
318         asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,        \
319                                           PARAVIRT_TYPE(type))  \
320                       : "=r"(ia64_clobber)                      \
321                       : PARAVIRT_OP(op), "0"(__##arg1)          \
322                       : PARAVIRT_CLOBBERS1)
323
324 #define PARAVIRT_BR1_RET(op, type, arg1)                        \
325         register unsigned long ia64_intri_res asm ("r8");       \
326         register unsigned long __##arg1 asm ("r8") = arg1;      \
327         asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,        \
328                                           PARAVIRT_TYPE(type))  \
329                       : "=r"(ia64_intri_res)                    \
330                       : PARAVIRT_OP(op), "0"(__##arg1)          \
331                       : PARAVIRT_CLOBBERS1)
332
333 #define PARAVIRT_BR1_VOID(op, type, arg1)                       \
334         register void *__##arg1 asm ("r8") = arg1;              \
335         register unsigned long ia64_clobber asm ("r8");         \
336         asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,        \
337                                           PARAVIRT_TYPE(type))  \
338                       : "=r"(ia64_clobber)                      \
339                       : PARAVIRT_OP(op), "0"(__##arg1)          \
340                       : PARAVIRT_CLOBBERS1)
341
342 #define PARAVIRT_BR2(op, type, arg1, arg2)                              \
343         register unsigned long __##arg1 asm ("r8") = arg1;              \
344         register unsigned long __##arg2 asm ("r9") = arg2;              \
345         register unsigned long ia64_clobber1 asm ("r8");                \
346         register unsigned long ia64_clobber2 asm ("r9");                \
347         asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,                \
348                                           PARAVIRT_TYPE(type))          \
349                       : "=r"(ia64_clobber1), "=r"(ia64_clobber2)        \
350                       : PARAVIRT_OP(op), "0"(__##arg1), "1"(__##arg2)   \
351                       : PARAVIRT_CLOBBERS2)
352
353
354 #define PARAVIRT_DEFINE_CPU_OP0(op, type)               \
355         static inline void                              \
356         paravirt_ ## op (void)                          \
357         {                                               \
358                 PARAVIRT_BR0(op, type);                 \
359         }
360
361 #define PARAVIRT_DEFINE_CPU_OP0_RET(op, type)           \
362         static inline unsigned long                     \
363         paravirt_ ## op (void)                          \
364         {                                               \
365                 PARAVIRT_BR0_RET(op, type);             \
366                 return ia64_intri_res;                  \
367         }
368
369 #define PARAVIRT_DEFINE_CPU_OP1_VOID(op, type)          \
370         static inline void                              \
371         paravirt_ ## op (void *arg1)                    \
372         {                                               \
373                 PARAVIRT_BR1_VOID(op, type, arg1);      \
374         }
375
376 #define PARAVIRT_DEFINE_CPU_OP1(op, type)               \
377         static inline void                              \
378         paravirt_ ## op (unsigned long arg1)            \
379         {                                               \
380                 PARAVIRT_BR1(op, type, arg1);           \
381         }
382
383 #define PARAVIRT_DEFINE_CPU_OP1_RET(op, type)           \
384         static inline unsigned long                     \
385         paravirt_ ## op (unsigned long arg1)            \
386         {                                               \
387                 PARAVIRT_BR1_RET(op, type, arg1);       \
388                 return ia64_intri_res;                  \
389         }
390
391 #define PARAVIRT_DEFINE_CPU_OP2(op, type)               \
392         static inline void                              \
393         paravirt_ ## op (unsigned long arg1,            \
394                          unsigned long arg2)            \
395         {                                               \
396                 PARAVIRT_BR2(op, type, arg1, arg2);     \
397         }
398
399
400 PARAVIRT_DEFINE_CPU_OP1_VOID(fc, FC);
401 PARAVIRT_DEFINE_CPU_OP1_RET(thash, THASH)
402 PARAVIRT_DEFINE_CPU_OP1_RET(get_cpuid, GET_CPUID)
403 PARAVIRT_DEFINE_CPU_OP1_RET(get_pmd, GET_PMD)
404 PARAVIRT_DEFINE_CPU_OP2(ptcga, PTCGA)
405 PARAVIRT_DEFINE_CPU_OP1_RET(get_rr, GET_RR)
406 PARAVIRT_DEFINE_CPU_OP2(set_rr, SET_RR)
407 PARAVIRT_DEFINE_CPU_OP0(ssm_i, SSM_I)
408 PARAVIRT_DEFINE_CPU_OP0(rsm_i, RSM_I)
409 PARAVIRT_DEFINE_CPU_OP0_RET(get_psr_i, GET_PSR_I)
410 PARAVIRT_DEFINE_CPU_OP1(intrin_local_irq_restore, INTRIN_LOCAL_IRQ_RESTORE)
411
412 static inline void
413 paravirt_set_rr0_to_rr4(unsigned long val0, unsigned long val1,
414                         unsigned long val2, unsigned long val3,
415                         unsigned long val4)
416 {
417         register unsigned long __val0 asm ("r8") = val0;
418         register unsigned long __val1 asm ("r9") = val1;
419         register unsigned long __val2 asm ("r10") = val2;
420         register unsigned long __val3 asm ("r11") = val3;
421         register unsigned long __val4 asm ("r14") = val4;
422
423         register unsigned long ia64_clobber0 asm ("r8");
424         register unsigned long ia64_clobber1 asm ("r9");
425         register unsigned long ia64_clobber2 asm ("r10");
426         register unsigned long ia64_clobber3 asm ("r11");
427         register unsigned long ia64_clobber4 asm ("r14");
428
429         asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,
430                                           PARAVIRT_TYPE(SET_RR0_TO_RR4))
431                       : "=r"(ia64_clobber0),
432                         "=r"(ia64_clobber1),
433                         "=r"(ia64_clobber2),
434                         "=r"(ia64_clobber3),
435                         "=r"(ia64_clobber4)
436                       : PARAVIRT_OP(set_rr0_to_rr4),
437                         "0"(__val0), "1"(__val1), "2"(__val2),
438                         "3"(__val3), "4"(__val4)
439                       : PARAVIRT_CLOBBERS5);
440 }
441
442 /* unsigned long paravirt_getreg(int reg) */
443 #define __paravirt_getreg(reg)                                          \
444         ({                                                              \
445                 register unsigned long ia64_intri_res asm ("r8");       \
446                 register unsigned long __reg asm ("r8") = (reg);        \
447                                                                         \
448                 asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,        \
449                                                   PARAVIRT_TYPE(GETREG) \
450                                                   + (reg))              \
451                               : "=r"(ia64_intri_res)                    \
452                               : PARAVIRT_OP(getreg), "0"(__reg)         \
453                               : PARAVIRT_CLOBBERS1);                    \
454                                                                         \
455                 ia64_intri_res;                                         \
456         })
457
458 /* void paravirt_setreg(int reg, unsigned long val) */
459 #define paravirt_setreg(reg, val)                                       \
460         do {                                                            \
461                 register unsigned long __val asm ("r8") = val;          \
462                 register unsigned long __reg asm ("r9") = reg;          \
463                 register unsigned long ia64_clobber1 asm ("r8");        \
464                 register unsigned long ia64_clobber2 asm ("r9");        \
465                                                                         \
466                 asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,        \
467                                                   PARAVIRT_TYPE(SETREG) \
468                                                   + (reg))              \
469                               : "=r"(ia64_clobber1),                    \
470                                 "=r"(ia64_clobber2)                     \
471                               : PARAVIRT_OP(setreg),                    \
472                                 "1"(__reg), "0"(__val)                  \
473                               : PARAVIRT_CLOBBERS2);                    \
474         } while (0)
475
476 #endif /* ASM_SUPPORTED */
477 #endif /* CONFIG_PARAVIRT && ASM_SUPPOTED */
478
479 #endif /* _ASM_IA64_PARAVIRT_PRIVOP_H */