158 lines
		
	
	
		
			4.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			158 lines
		
	
	
		
			4.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* SPDX-License-Identifier: GPL-2.0 */
 | |
| #ifndef _LINUX_KERNEL_VTIME_H
 | |
| #define _LINUX_KERNEL_VTIME_H
 | |
| 
 | |
| #include <linux/context_tracking_state.h>
 | |
| #include <linux/sched.h>
 | |
| 
 | |
| /*
 | |
|  * Common vtime APIs
 | |
|  */
 | |
| #ifdef CONFIG_VIRT_CPU_ACCOUNTING
 | |
| extern void vtime_account_kernel(struct task_struct *tsk);
 | |
| extern void vtime_account_idle(struct task_struct *tsk);
 | |
| #endif /* !CONFIG_VIRT_CPU_ACCOUNTING */
 | |
| 
 | |
| #ifdef CONFIG_VIRT_CPU_ACCOUNTING_GEN
 | |
| extern void vtime_user_enter(struct task_struct *tsk);
 | |
| extern void vtime_user_exit(struct task_struct *tsk);
 | |
| extern void vtime_guest_enter(struct task_struct *tsk);
 | |
| extern void vtime_guest_exit(struct task_struct *tsk);
 | |
| extern void vtime_init_idle(struct task_struct *tsk, int cpu);
 | |
| #else /* !CONFIG_VIRT_CPU_ACCOUNTING_GEN  */
 | |
| static inline void vtime_user_enter(struct task_struct *tsk) { }
 | |
| static inline void vtime_user_exit(struct task_struct *tsk) { }
 | |
| static inline void vtime_guest_enter(struct task_struct *tsk) { }
 | |
| static inline void vtime_guest_exit(struct task_struct *tsk) { }
 | |
| static inline void vtime_init_idle(struct task_struct *tsk, int cpu) { }
 | |
| #endif
 | |
| 
 | |
| #ifdef CONFIG_VIRT_CPU_ACCOUNTING_NATIVE
 | |
| extern void vtime_account_irq(struct task_struct *tsk, unsigned int offset);
 | |
| extern void vtime_account_softirq(struct task_struct *tsk);
 | |
| extern void vtime_account_hardirq(struct task_struct *tsk);
 | |
| extern void vtime_flush(struct task_struct *tsk);
 | |
| #else /* !CONFIG_VIRT_CPU_ACCOUNTING_NATIVE */
 | |
| static inline void vtime_account_irq(struct task_struct *tsk, unsigned int offset) { }
 | |
| static inline void vtime_account_softirq(struct task_struct *tsk) { }
 | |
| static inline void vtime_account_hardirq(struct task_struct *tsk) { }
 | |
| static inline void vtime_flush(struct task_struct *tsk) { }
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * vtime_accounting_enabled_this_cpu() definitions/declarations
 | |
|  */
 | |
| #if defined(CONFIG_VIRT_CPU_ACCOUNTING_NATIVE)
 | |
| 
 | |
| static inline bool vtime_accounting_enabled_this_cpu(void) { return true; }
 | |
| extern void vtime_task_switch(struct task_struct *prev);
 | |
| 
 | |
| static __always_inline void vtime_account_guest_enter(void)
 | |
| {
 | |
| 	vtime_account_kernel(current);
 | |
| 	current->flags |= PF_VCPU;
 | |
| }
 | |
| 
 | |
| static __always_inline void vtime_account_guest_exit(void)
 | |
| {
 | |
| 	vtime_account_kernel(current);
 | |
| 	current->flags &= ~PF_VCPU;
 | |
| }
 | |
| 
 | |
| #elif defined(CONFIG_VIRT_CPU_ACCOUNTING_GEN)
 | |
| 
 | |
| /*
 | |
|  * Checks if vtime is enabled on some CPU. Cputime readers want to be careful
 | |
|  * in that case and compute the tickless cputime.
 | |
|  * For now vtime state is tied to context tracking. We might want to decouple
 | |
|  * those later if necessary.
 | |
|  */
 | |
| static inline bool vtime_accounting_enabled(void)
 | |
| {
 | |
| 	return context_tracking_enabled();
 | |
| }
 | |
| 
 | |
| static inline bool vtime_accounting_enabled_cpu(int cpu)
 | |
| {
 | |
| 	return context_tracking_enabled_cpu(cpu);
 | |
| }
 | |
| 
 | |
| static inline bool vtime_accounting_enabled_this_cpu(void)
 | |
| {
 | |
| 	return context_tracking_enabled_this_cpu();
 | |
| }
 | |
| 
 | |
| extern void vtime_task_switch_generic(struct task_struct *prev);
 | |
| 
 | |
| static inline void vtime_task_switch(struct task_struct *prev)
 | |
| {
 | |
| 	if (vtime_accounting_enabled_this_cpu())
 | |
| 		vtime_task_switch_generic(prev);
 | |
| }
 | |
| 
 | |
| static __always_inline void vtime_account_guest_enter(void)
 | |
| {
 | |
| 	if (vtime_accounting_enabled_this_cpu())
 | |
| 		vtime_guest_enter(current);
 | |
| 	else
 | |
| 		current->flags |= PF_VCPU;
 | |
| }
 | |
| 
 | |
| static __always_inline void vtime_account_guest_exit(void)
 | |
| {
 | |
| 	if (vtime_accounting_enabled_this_cpu())
 | |
| 		vtime_guest_exit(current);
 | |
| 	else
 | |
| 		current->flags &= ~PF_VCPU;
 | |
| }
 | |
| 
 | |
| #else /* !CONFIG_VIRT_CPU_ACCOUNTING */
 | |
| 
 | |
| static inline bool vtime_accounting_enabled_this_cpu(void) { return false; }
 | |
| static inline void vtime_task_switch(struct task_struct *prev) { }
 | |
| 
 | |
| static __always_inline void vtime_account_guest_enter(void)
 | |
| {
 | |
| 	current->flags |= PF_VCPU;
 | |
| }
 | |
| 
 | |
| static __always_inline void vtime_account_guest_exit(void)
 | |
| {
 | |
| 	current->flags &= ~PF_VCPU;
 | |
| }
 | |
| 
 | |
| #endif
 | |
| 
 | |
| 
 | |
| #ifdef CONFIG_IRQ_TIME_ACCOUNTING
 | |
| extern void irqtime_account_irq(struct task_struct *tsk, unsigned int offset);
 | |
| #else
 | |
| static inline void irqtime_account_irq(struct task_struct *tsk, unsigned int offset) { }
 | |
| #endif
 | |
| 
 | |
| static inline void account_softirq_enter(struct task_struct *tsk)
 | |
| {
 | |
| 	vtime_account_irq(tsk, SOFTIRQ_OFFSET);
 | |
| 	irqtime_account_irq(tsk, SOFTIRQ_OFFSET);
 | |
| }
 | |
| 
 | |
| static inline void account_softirq_exit(struct task_struct *tsk)
 | |
| {
 | |
| 	vtime_account_softirq(tsk);
 | |
| 	irqtime_account_irq(tsk, 0);
 | |
| }
 | |
| 
 | |
| static inline void account_hardirq_enter(struct task_struct *tsk)
 | |
| {
 | |
| 	vtime_account_irq(tsk, HARDIRQ_OFFSET);
 | |
| 	irqtime_account_irq(tsk, HARDIRQ_OFFSET);
 | |
| }
 | |
| 
 | |
| static inline void account_hardirq_exit(struct task_struct *tsk)
 | |
| {
 | |
| 	vtime_account_hardirq(tsk);
 | |
| 	irqtime_account_irq(tsk, 0);
 | |
| }
 | |
| 
 | |
| #endif /* _LINUX_KERNEL_VTIME_H */
 |