aboutsummaryrefslogtreecommitdiff
path: root/include/linux/mt_sched_mon.h
diff options
context:
space:
mode:
authorMeizu OpenSource <patchwork@meizu.com>2016-08-15 10:19:42 +0800
committerMeizu OpenSource <patchwork@meizu.com>2016-08-15 10:19:42 +0800
commitd2e1446d81725c351dc73a03b397ce043fb18452 (patch)
tree4dbc616b7f92aea39cd697a9084205ddb805e344 /include/linux/mt_sched_mon.h
downloadandroid_kernel_m2note-d2e1446d81725c351dc73a03b397ce043fb18452.tar.gz
first commit
Diffstat (limited to 'include/linux/mt_sched_mon.h')
-rw-r--r--include/linux/mt_sched_mon.h80
1 files changed, 80 insertions, 0 deletions
diff --git a/include/linux/mt_sched_mon.h b/include/linux/mt_sched_mon.h
new file mode 100644
index 000000000..e3c4f93f1
--- /dev/null
+++ b/include/linux/mt_sched_mon.h
@@ -0,0 +1,80 @@
+/*CPU holding event: ISR/SoftIRQ/Tasklet/Timer*/
+struct sched_block_event {
+ int type;
+ unsigned long cur_event;
+ unsigned long last_event;
+ unsigned long long cur_ts;
+ unsigned long long last_ts;
+ unsigned long long last_te;
+
+ unsigned long cur_count;
+ unsigned long last_count;
+
+ int preempt_count;
+};
+DECLARE_PER_CPU(struct sched_block_event, ISR_mon);
+DECLARE_PER_CPU(struct sched_block_event, SoftIRQ_mon);
+DECLARE_PER_CPU(struct sched_block_event, tasklet_mon);
+DECLARE_PER_CPU(struct sched_block_event, hrt_mon);
+DECLARE_PER_CPU(struct sched_block_event, sft_mon);
+
+DECLARE_PER_CPU(int, mt_timer_irq);
+extern void mt_trace_ISR_start(int id);
+extern void mt_trace_ISR_end(int id);
+extern void mt_trace_SoftIRQ_start(int id);
+extern void mt_trace_SoftIRQ_end(int id);
+extern void mt_trace_tasklet_start(void *func);
+extern void mt_trace_tasklet_end(void *func);
+extern void mt_trace_hrt_start(void *func);
+extern void mt_trace_hrt_end(void *func);
+extern void mt_trace_sft_start(void *func);
+extern void mt_trace_sft_end(void *func);
+
+extern void mt_save_irq_counts(void);
+extern void mt_show_last_irq_counts(void);
+extern void mt_show_current_irq_counts(void);
+
+/*Schedule disable event: IRQ/Preempt disable monitor*/
+struct sched_stop_event {
+ unsigned long long cur_ts;
+ unsigned long long last_ts;
+ unsigned long long last_te;
+};
+DECLARE_PER_CPU(struct sched_stop_event, IRQ_disable_mon);
+DECLARE_PER_CPU(struct sched_stop_event, Preempt_disable_mon);
+extern void MT_trace_irq_on(void);
+extern void MT_trace_irq_off(void);
+extern void MT_trace_preempt_on(void);
+extern void MT_trace_preempt_off(void);
+/* [IRQ-disable] White List
+ * Flags for special scenario*/
+DECLARE_PER_CPU(int, MT_trace_in_sched);
+DECLARE_PER_CPU(int, MT_trace_in_resume_console);
+
+extern void mt_aee_dump_sched_traces(void);
+extern void mt_dump_sched_traces(void);
+
+DECLARE_PER_CPU(int, mtsched_mon_enabled);
+DECLARE_PER_CPU(unsigned long long, local_timer_ts);
+DECLARE_PER_CPU(unsigned long long, local_timer_te);
+
+#define MT_LOCAL_TIMER_IRQ 29
+#define MT_LOCAL_WDT_IRQ 3
+#define MAX_NR_IRQS 512
+
+#ifdef CONFIG_MT_SCHED_MONITOR
+#define mt_sched_mon_disable(sflag) \
+do { \
+ sflag = __raw_get_cpu_var(mtsched_mon_enabled); \
+ __raw_get_cpu_var(mtsched_mon_enabled) = 0; \
+} while (0)
+
+#define mt_sched_mon_restore(sflag) \
+do { \
+ __raw_get_cpu_var(mtsched_mon_enabled) = sflag; \
+} while (0)
+#else
+
+#define mt_sched_mon_disable(sflag) do {} while (0)
+#define mt_sched_mon_restore(sflag) do {} while (0)
+#endif