1212 lines
		
	
	
		
			30 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
			
		
		
	
	
			1212 lines
		
	
	
		
			30 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
From d001bd8483c805c45a42d9bd0468a96722e72875 Mon Sep 17 00:00:00 2001
 | 
						|
From: Grissiom <chaos.proton@gmail.com>
 | 
						|
Date: Thu, 1 Aug 2013 14:59:56 +0800
 | 
						|
Subject: [PATCH 1/2] RTT-VMM: implement dual system running on realview-pb-a8
 | 
						|
 | 
						|
Signed-off-by: Grissiom <chaos.proton@gmail.com>
 | 
						|
Signed-off-by: Bernard.Xiong <bernard.xiong@gmail.com>
 | 
						|
---
 | 
						|
 arch/arm/Kconfig                   |   1 +
 | 
						|
 arch/arm/Makefile                  |   1 +
 | 
						|
 arch/arm/common/gic.c              |  67 +++++++++++++-
 | 
						|
 arch/arm/include/asm/assembler.h   |   8 +-
 | 
						|
 arch/arm/include/asm/domain.h      |   7 ++
 | 
						|
 arch/arm/include/asm/irqflags.h    |  84 ++++++++++++-----
 | 
						|
 arch/arm/include/asm/mach/map.h    |   5 +
 | 
						|
 arch/arm/include/vmm/vmm.h         |  35 +++++++
 | 
						|
 arch/arm/include/vmm/vmm_config.h  |   7 ++
 | 
						|
 arch/arm/kernel/entry-armv.S       |  30 +++++-
 | 
						|
 arch/arm/kernel/entry-common.S     |   3 +
 | 
						|
 arch/arm/kernel/entry-header.S     |  15 ++-
 | 
						|
 arch/arm/mach-omap2/irq.c          |  12 +++
 | 
						|
 arch/arm/mm/fault.c                |   9 ++
 | 
						|
 arch/arm/mm/init.c                 |   8 ++
 | 
						|
 arch/arm/mm/mmu.c                  |  44 +++++++++
 | 
						|
 arch/arm/vmm/Kconfig               |  49 ++++++++++
 | 
						|
 arch/arm/vmm/Makefile              |  10 ++
 | 
						|
 arch/arm/vmm/README                |   1 +
 | 
						|
 arch/arm/vmm/am33xx/intc.h         |  13 +++
 | 
						|
 arch/arm/vmm/am33xx/softirq.c      |  14 +++
 | 
						|
 arch/arm/vmm/am33xx/virq.c         |  48 ++++++++++
 | 
						|
 arch/arm/vmm/realview_a8/softirq.c |  12 +++
 | 
						|
 arch/arm/vmm/vmm.c                 |  32 +++++++
 | 
						|
 arch/arm/vmm/vmm_traps.c           |  37 ++++++++
 | 
						|
 arch/arm/vmm/vmm_virhw.h           |  59 ++++++++++++
 | 
						|
 arch/arm/vmm/vmm_virq.c            | 183 +++++++++++++++++++++++++++++++++++++
 | 
						|
 27 files changed, 767 insertions(+), 27 deletions(-)
 | 
						|
 create mode 100644 arch/arm/include/vmm/vmm.h
 | 
						|
 create mode 100644 arch/arm/include/vmm/vmm_config.h
 | 
						|
 create mode 100644 arch/arm/vmm/Kconfig
 | 
						|
 create mode 100644 arch/arm/vmm/Makefile
 | 
						|
 create mode 100644 arch/arm/vmm/README
 | 
						|
 create mode 100644 arch/arm/vmm/am33xx/intc.h
 | 
						|
 create mode 100644 arch/arm/vmm/am33xx/softirq.c
 | 
						|
 create mode 100644 arch/arm/vmm/am33xx/virq.c
 | 
						|
 create mode 100644 arch/arm/vmm/realview_a8/softirq.c
 | 
						|
 create mode 100644 arch/arm/vmm/vmm.c
 | 
						|
 create mode 100644 arch/arm/vmm/vmm_traps.c
 | 
						|
 create mode 100644 arch/arm/vmm/vmm_virhw.h
 | 
						|
 create mode 100644 arch/arm/vmm/vmm_virq.c
 | 
						|
 | 
						|
diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig
 | 
						|
index 67874b8..eb82cd6 100644
 | 
						|
--- a/arch/arm/Kconfig
 | 
						|
+++ b/arch/arm/Kconfig
 | 
						|
@@ -1164,6 +1164,7 @@ config ARM_TIMER_SP804
 | 
						|
 	select HAVE_SCHED_CLOCK
 | 
						|
 
 | 
						|
 source arch/arm/mm/Kconfig
 | 
						|
+source arch/arm/vmm/Kconfig
 | 
						|
 
 | 
						|
 config ARM_NR_BANKS
 | 
						|
 	int
 | 
						|
diff --git a/arch/arm/Makefile b/arch/arm/Makefile
 | 
						|
index 30c443c..262c8e2 100644
 | 
						|
--- a/arch/arm/Makefile
 | 
						|
+++ b/arch/arm/Makefile
 | 
						|
@@ -252,6 +252,7 @@ core-$(CONFIG_FPE_NWFPE)	+= arch/arm/nwfpe/
 | 
						|
 core-$(CONFIG_FPE_FASTFPE)	+= $(FASTFPE_OBJ)
 | 
						|
 core-$(CONFIG_VFP)		+= arch/arm/vfp/
 | 
						|
 core-$(CONFIG_XEN)		+= arch/arm/xen/
 | 
						|
+core-$(CONFIG_ARM_VMM)		+= arch/arm/vmm/
 | 
						|
 
 | 
						|
 # If we have a machine-specific directory, then include it in the build.
 | 
						|
 core-y				+= arch/arm/kernel/ arch/arm/mm/ arch/arm/common/
 | 
						|
diff --git a/arch/arm/common/gic.c b/arch/arm/common/gic.c
 | 
						|
index 87dfa90..a9d7357 100644
 | 
						|
--- a/arch/arm/common/gic.c
 | 
						|
+++ b/arch/arm/common/gic.c
 | 
						|
@@ -45,6 +45,11 @@
 | 
						|
 #include <asm/mach/irq.h>
 | 
						|
 #include <asm/hardware/gic.h>
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+#include "../vmm/vmm_virhw.h"
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 union gic_base {
 | 
						|
 	void __iomem *common_base;
 | 
						|
 	void __percpu __iomem **percpu_base;
 | 
						|
@@ -276,12 +281,72 @@ static int gic_set_wake(struct irq_data *d, unsigned int on)
 | 
						|
 #define gic_set_wake	NULL
 | 
						|
 #endif
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+void vmm_irq_handle(struct gic_chip_data *gic, struct pt_regs *regs)
 | 
						|
+{
 | 
						|
+	unsigned long flags;
 | 
						|
+	struct vmm_context* _vmm_context;
 | 
						|
+
 | 
						|
+	_vmm_context = vmm_context_get();
 | 
						|
+
 | 
						|
+	while (_vmm_context->virq_pended) {
 | 
						|
+		int index;
 | 
						|
+
 | 
						|
+		flags = vmm_irq_save();
 | 
						|
+		_vmm_context->virq_pended = 0;
 | 
						|
+		vmm_irq_restore(flags);
 | 
						|
+
 | 
						|
+		/* get the pending interrupt */
 | 
						|
+		for (index = 0; index < IRQS_NR_32; index++) {
 | 
						|
+			int pdbit;
 | 
						|
+
 | 
						|
+			for (pdbit = __builtin_ffs(_vmm_context->virq_pending[index]);
 | 
						|
+			     pdbit != 0;
 | 
						|
+			     pdbit = __builtin_ffs(_vmm_context->virq_pending[index])) {
 | 
						|
+				unsigned long inner_flag;
 | 
						|
+				int irqnr, oirqnr;
 | 
						|
+
 | 
						|
+				pdbit--;
 | 
						|
+
 | 
						|
+				inner_flag = vmm_irq_save();
 | 
						|
+				_vmm_context->virq_pending[index] &= ~(1 << pdbit);
 | 
						|
+				vmm_irq_restore(inner_flag);
 | 
						|
+
 | 
						|
+				oirqnr = pdbit + index * 32;
 | 
						|
+				if (likely(oirqnr > 15 && oirqnr < 1021)) {
 | 
						|
+					irqnr = irq_find_mapping(gic->domain, oirqnr);
 | 
						|
+					handle_IRQ(irqnr, regs);
 | 
						|
+				} else if (oirqnr < 16) {
 | 
						|
+					/* soft IRQs are EOIed by the host. */
 | 
						|
+#ifdef CONFIG_SMP
 | 
						|
+					handle_IPI(oirqnr, regs);
 | 
						|
+#endif
 | 
						|
+				}
 | 
						|
+				/* umask interrupt */
 | 
						|
+				/* FIXME: maybe we don't need this */
 | 
						|
+				writel_relaxed(1 << (oirqnr % 32),
 | 
						|
+					       gic_data_dist_base(gic)
 | 
						|
+					       + GIC_DIST_ENABLE_SET
 | 
						|
+					       + (oirqnr / 32) * 4);
 | 
						|
+
 | 
						|
+			}
 | 
						|
+		}
 | 
						|
+	}
 | 
						|
+}
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 asmlinkage void __exception_irq_entry gic_handle_irq(struct pt_regs *regs)
 | 
						|
 {
 | 
						|
 	u32 irqstat, irqnr;
 | 
						|
 	struct gic_chip_data *gic = &gic_data[0];
 | 
						|
 	void __iomem *cpu_base = gic_data_cpu_base(gic);
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+	if (vmm_get_status()) {
 | 
						|
+		vmm_irq_handle(gic, regs);
 | 
						|
+		return;
 | 
						|
+	}
 | 
						|
+#endif
 | 
						|
 	do {
 | 
						|
 		irqstat = readl_relaxed(cpu_base + GIC_CPU_INTACK);
 | 
						|
 		irqnr = irqstat & ~0x1c00;
 | 
						|
@@ -777,7 +842,7 @@ void __cpuinit gic_secondary_init(unsigned int gic_nr)
 | 
						|
 	gic_cpu_init(&gic_data[gic_nr]);
 | 
						|
 }
 | 
						|
 
 | 
						|
-#ifdef CONFIG_SMP
 | 
						|
+#if defined(CONFIG_SMP) || defined(CONFIG_ARM_VMM)
 | 
						|
 void gic_raise_softirq(const struct cpumask *mask, unsigned int irq)
 | 
						|
 {
 | 
						|
 	int cpu;
 | 
						|
diff --git a/arch/arm/include/asm/assembler.h b/arch/arm/include/asm/assembler.h
 | 
						|
index eb87200..b646fa7 100644
 | 
						|
--- a/arch/arm/include/asm/assembler.h
 | 
						|
+++ b/arch/arm/include/asm/assembler.h
 | 
						|
@@ -82,11 +82,15 @@
 | 
						|
  */
 | 
						|
 #if __LINUX_ARM_ARCH__ >= 6
 | 
						|
 	.macro	disable_irq_notrace
 | 
						|
-	cpsid	i
 | 
						|
+	stmdb   sp!, {r0-r3, ip, lr}
 | 
						|
+	bl	irq_disable_asm
 | 
						|
+	ldmia	sp!, {r0-r3, ip, lr}
 | 
						|
 	.endm
 | 
						|
 
 | 
						|
 	.macro	enable_irq_notrace
 | 
						|
-	cpsie	i
 | 
						|
+	stmdb   sp!, {r0-r3, ip, lr}
 | 
						|
+	bl	irq_enable_asm
 | 
						|
+	ldmia	sp!, {r0-r3, ip, lr}
 | 
						|
 	.endm
 | 
						|
 #else
 | 
						|
 	.macro	disable_irq_notrace
 | 
						|
diff --git a/arch/arm/include/asm/domain.h b/arch/arm/include/asm/domain.h
 | 
						|
index 6ddbe44..bbc4470 100644
 | 
						|
--- a/arch/arm/include/asm/domain.h
 | 
						|
+++ b/arch/arm/include/asm/domain.h
 | 
						|
@@ -44,6 +44,13 @@
 | 
						|
 #define DOMAIN_IO	0
 | 
						|
 #endif
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+/* RT-Thread VMM memory space */
 | 
						|
+#define DOMAIN_RTVMM      3
 | 
						|
+/* shared memory with VMM and Linux */
 | 
						|
+#define DOMAIN_RTVMM_SHR  4
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 /*
 | 
						|
  * Domain types
 | 
						|
  */
 | 
						|
diff --git a/arch/arm/include/asm/irqflags.h b/arch/arm/include/asm/irqflags.h
 | 
						|
index 1e6cca5..bfaedff 100644
 | 
						|
--- a/arch/arm/include/asm/irqflags.h
 | 
						|
+++ b/arch/arm/include/asm/irqflags.h
 | 
						|
@@ -9,34 +9,56 @@
 | 
						|
  * CPU interrupt mask handling.
 | 
						|
  */
 | 
						|
 #if __LINUX_ARM_ARCH__ >= 6
 | 
						|
+#include <vmm/vmm.h> /* VMM only support ARMv7 right now */
 | 
						|
 
 | 
						|
 static inline unsigned long arch_local_irq_save(void)
 | 
						|
 {
 | 
						|
 	unsigned long flags;
 | 
						|
 
 | 
						|
-	asm volatile(
 | 
						|
-		"	mrs	%0, cpsr	@ arch_local_irq_save\n"
 | 
						|
-		"	cpsid	i"
 | 
						|
-		: "=r" (flags) : : "memory", "cc");
 | 
						|
+	if (vmm_status)
 | 
						|
+	{
 | 
						|
+		flags = vmm_save_virq();
 | 
						|
+	}
 | 
						|
+	else
 | 
						|
+	{
 | 
						|
+		asm volatile(
 | 
						|
+			"	mrs	%0, cpsr	@ arch_local_irq_save\n"
 | 
						|
+			"	cpsid	i"
 | 
						|
+			: "=r" (flags) : : "memory", "cc");
 | 
						|
+	}
 | 
						|
 	return flags;
 | 
						|
 }
 | 
						|
 
 | 
						|
 static inline void arch_local_irq_enable(void)
 | 
						|
 {
 | 
						|
-	asm volatile(
 | 
						|
-		"	cpsie i			@ arch_local_irq_enable"
 | 
						|
-		:
 | 
						|
-		:
 | 
						|
-		: "memory", "cc");
 | 
						|
+	if (vmm_status)
 | 
						|
+	{
 | 
						|
+		vmm_enable_virq();
 | 
						|
+	}
 | 
						|
+	else
 | 
						|
+	{
 | 
						|
+		asm volatile(
 | 
						|
+			"	cpsie i			@ arch_local_irq_enable"
 | 
						|
+			:
 | 
						|
+			:
 | 
						|
+			: "memory", "cc");
 | 
						|
+	}
 | 
						|
 }
 | 
						|
 
 | 
						|
 static inline void arch_local_irq_disable(void)
 | 
						|
 {
 | 
						|
-	asm volatile(
 | 
						|
-		"	cpsid i			@ arch_local_irq_disable"
 | 
						|
-		:
 | 
						|
-		:
 | 
						|
-		: "memory", "cc");
 | 
						|
+	if (vmm_status)
 | 
						|
+	{
 | 
						|
+		vmm_disable_virq();
 | 
						|
+	}
 | 
						|
+	else
 | 
						|
+	{
 | 
						|
+		asm volatile(
 | 
						|
+			"	cpsid i			@ arch_local_irq_disable"
 | 
						|
+			:
 | 
						|
+			:
 | 
						|
+			: "memory", "cc");
 | 
						|
+	}
 | 
						|
 }
 | 
						|
 
 | 
						|
 #define local_fiq_enable()  __asm__("cpsie f	@ __stf" : : : "memory", "cc")
 | 
						|
@@ -128,9 +150,17 @@ static inline void arch_local_irq_disable(void)
 | 
						|
 static inline unsigned long arch_local_save_flags(void)
 | 
						|
 {
 | 
						|
 	unsigned long flags;
 | 
						|
-	asm volatile(
 | 
						|
-		"	mrs	%0, cpsr	@ local_save_flags"
 | 
						|
-		: "=r" (flags) : : "memory", "cc");
 | 
						|
+
 | 
						|
+	if (vmm_status)
 | 
						|
+	{
 | 
						|
+		flags = vmm_return_virq();
 | 
						|
+	}
 | 
						|
+	else
 | 
						|
+	{
 | 
						|
+		asm volatile(
 | 
						|
+			"	mrs	%0, cpsr	@ local_save_flags"
 | 
						|
+			: "=r" (flags) : : "memory", "cc");
 | 
						|
+	}
 | 
						|
 	return flags;
 | 
						|
 }
 | 
						|
 
 | 
						|
@@ -139,15 +169,25 @@ static inline unsigned long arch_local_save_flags(void)
 | 
						|
  */
 | 
						|
 static inline void arch_local_irq_restore(unsigned long flags)
 | 
						|
 {
 | 
						|
-	asm volatile(
 | 
						|
-		"	msr	cpsr_c, %0	@ local_irq_restore"
 | 
						|
-		:
 | 
						|
-		: "r" (flags)
 | 
						|
-		: "memory", "cc");
 | 
						|
+	if (vmm_status)
 | 
						|
+	{
 | 
						|
+		vmm_restore_virq(flags);
 | 
						|
+	}
 | 
						|
+	else
 | 
						|
+	{
 | 
						|
+		asm volatile(
 | 
						|
+			"	msr	cpsr_c, %0	@ local_irq_restore"
 | 
						|
+			:
 | 
						|
+			: "r" (flags)
 | 
						|
+			: "memory", "cc");
 | 
						|
+	}
 | 
						|
 }
 | 
						|
 
 | 
						|
 static inline int arch_irqs_disabled_flags(unsigned long flags)
 | 
						|
 {
 | 
						|
+	if (vmm_status)
 | 
						|
+		return (flags == 0x01);
 | 
						|
+
 | 
						|
 	return flags & PSR_I_BIT;
 | 
						|
 }
 | 
						|
 
 | 
						|
diff --git a/arch/arm/include/asm/mach/map.h b/arch/arm/include/asm/mach/map.h
 | 
						|
index 2fe141f..502b341 100644
 | 
						|
--- a/arch/arm/include/asm/mach/map.h
 | 
						|
+++ b/arch/arm/include/asm/mach/map.h
 | 
						|
@@ -35,6 +35,11 @@ struct map_desc {
 | 
						|
 #define MT_MEMORY_SO		14
 | 
						|
 #define MT_MEMORY_DMA_READY	15
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+#define MT_RTVMM                16
 | 
						|
+#define MT_RTVMM_SHARE          17
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 #ifdef CONFIG_MMU
 | 
						|
 extern void iotable_init(struct map_desc *, int);
 | 
						|
 extern void vm_reserve_area_early(unsigned long addr, unsigned long size,
 | 
						|
diff --git a/arch/arm/include/vmm/vmm.h b/arch/arm/include/vmm/vmm.h
 | 
						|
new file mode 100644
 | 
						|
index 0000000..3ff3f31
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/include/vmm/vmm.h
 | 
						|
@@ -0,0 +1,35 @@
 | 
						|
+#ifndef __LINUX_VMM_H__
 | 
						|
+#define __LINUX_VMM_H__
 | 
						|
+
 | 
						|
+#include <linux/compiler.h>
 | 
						|
+
 | 
						|
+#include "vmm_config.h"
 | 
						|
+
 | 
						|
+struct irq_domain;
 | 
						|
+struct pt_regs;
 | 
						|
+
 | 
						|
+extern int vmm_status;
 | 
						|
+extern struct vmm_context *_vmm_context;
 | 
						|
+
 | 
						|
+/* VMM context routines */
 | 
						|
+void vmm_context_init(void* context);
 | 
						|
+struct vmm_context* vmm_context_get(void);
 | 
						|
+
 | 
						|
+void vmm_set_status(int status);
 | 
						|
+int vmm_get_status(void);
 | 
						|
+
 | 
						|
+void vmm_mem_init(void);
 | 
						|
+void vmm_raise_softirq(int irq);
 | 
						|
+
 | 
						|
+/* VMM vIRQ routines */
 | 
						|
+unsigned long vmm_save_virq(void);
 | 
						|
+unsigned long vmm_return_virq(void);
 | 
						|
+
 | 
						|
+void vmm_restore_virq(unsigned long flags);
 | 
						|
+void vmm_enable_virq(void);
 | 
						|
+void vmm_disable_virq(void);
 | 
						|
+void vmm_enter_hw_noirq(void);
 | 
						|
+
 | 
						|
+void vmm_raise_softirq(int irq);
 | 
						|
+
 | 
						|
+#endif
 | 
						|
diff --git a/arch/arm/include/vmm/vmm_config.h b/arch/arm/include/vmm/vmm_config.h
 | 
						|
new file mode 100644
 | 
						|
index 0000000..cce5e8a
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/include/vmm/vmm_config.h
 | 
						|
@@ -0,0 +1,7 @@
 | 
						|
+#ifndef __LINUX_VMM_CONFIG_H__
 | 
						|
+#define __LINUX_VMM_CONFIG_H__
 | 
						|
+
 | 
						|
+#define HOST_VMM_ADDR_END 		CONFIG_HOST_VMM_ADDR_END
 | 
						|
+#define HOST_VMM_ADDR_BEGIN 	(CONFIG_HOST_VMM_ADDR_END - CONFIG_HOST_VMM_SIZE)
 | 
						|
+
 | 
						|
+#endif
 | 
						|
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S
 | 
						|
index 0f82098..80f1681 100644
 | 
						|
--- a/arch/arm/kernel/entry-armv.S
 | 
						|
+++ b/arch/arm/kernel/entry-armv.S
 | 
						|
@@ -182,6 +182,15 @@ ENDPROC(__und_invalid)
 | 
						|
 	@
 | 
						|
 	stmia	r7, {r2 - r6}
 | 
						|
 
 | 
						|
+	stmdb   sp!, {r0-r3, ip, lr}
 | 
						|
+	mov     r0, r5
 | 
						|
+	add     r1, sp, #4*6
 | 
						|
+	bl      vmm_save_virq_spsr_asm
 | 
						|
+	mov     r5, r0
 | 
						|
+	bl      vmm_switch_nohwirq_to_novirq
 | 
						|
+	ldmia   sp!, {r0-r3, ip, lr}
 | 
						|
+	str     r5, [sp, #S_PSR]              @ fix the pushed SPSR
 | 
						|
+
 | 
						|
 #ifdef CONFIG_TRACE_IRQFLAGS
 | 
						|
 	bl	trace_hardirqs_off
 | 
						|
 #endif
 | 
						|
@@ -208,6 +217,23 @@ __dabt_svc:
 | 
						|
  UNWIND(.fnend		)
 | 
						|
 ENDPROC(__dabt_svc)
 | 
						|
 
 | 
						|
+	.macro	svc_exit_irq, rpsr
 | 
						|
+	cpsid   i
 | 
						|
+	msr	spsr_cxsf, \rpsr
 | 
						|
+	mov     r0, \rpsr
 | 
						|
+	bl      vmm_on_svc_exit_irq
 | 
						|
+#if defined(CONFIG_CPU_V6)
 | 
						|
+	ldr	r0, [sp]
 | 
						|
+	strex	r1, r2, [sp]			@ clear the exclusive monitor
 | 
						|
+	ldmib	sp, {r1 - pc}^			@ load r1 - pc, cpsr
 | 
						|
+#elif defined(CONFIG_CPU_32v6K)
 | 
						|
+	clrex					@ clear the exclusive monitor
 | 
						|
+	ldmia	sp, {r0 - pc}^			@ load r0 - pc, cpsr
 | 
						|
+#else
 | 
						|
+	ldmia	sp, {r0 - pc}^			@ load r0 - pc, cpsr
 | 
						|
+#endif
 | 
						|
+	.endm
 | 
						|
+
 | 
						|
 	.align	5
 | 
						|
 __irq_svc:
 | 
						|
 	svc_entry
 | 
						|
@@ -228,7 +254,7 @@ __irq_svc:
 | 
						|
 	@ the first place, so there's no point checking the PSR I bit.
 | 
						|
 	bl	trace_hardirqs_on
 | 
						|
 #endif
 | 
						|
-	svc_exit r5				@ return from exception
 | 
						|
+	svc_exit_irq r5				@ return from exception
 | 
						|
  UNWIND(.fnend		)
 | 
						|
 ENDPROC(__irq_svc)
 | 
						|
 
 | 
						|
@@ -393,6 +419,8 @@ ENDPROC(__pabt_svc)
 | 
						|
 	@
 | 
						|
 	zero_fp
 | 
						|
 
 | 
						|
+	bl      vmm_switch_nohwirq_to_novirq
 | 
						|
+
 | 
						|
 #ifdef CONFIG_IRQSOFF_TRACER
 | 
						|
 	bl	trace_hardirqs_off
 | 
						|
 #endif
 | 
						|
diff --git a/arch/arm/kernel/entry-common.S b/arch/arm/kernel/entry-common.S
 | 
						|
index a6c301e..325a26e 100644
 | 
						|
--- a/arch/arm/kernel/entry-common.S
 | 
						|
+++ b/arch/arm/kernel/entry-common.S
 | 
						|
@@ -349,6 +349,9 @@ ENTRY(vector_swi)
 | 
						|
 	str	lr, [sp, #S_PC]			@ Save calling PC
 | 
						|
 	str	r8, [sp, #S_PSR]		@ Save CPSR
 | 
						|
 	str	r0, [sp, #S_OLD_R0]		@ Save OLD_R0
 | 
						|
+	stmdb   sp!, {r0-r3, ip, lr}
 | 
						|
+	bl	vmm_switch_nohwirq_to_novirq
 | 
						|
+	ldmia	sp!, {r0-r3, ip, lr}
 | 
						|
 	zero_fp
 | 
						|
 
 | 
						|
 	/*
 | 
						|
diff --git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S
 | 
						|
index 9a8531e..9e438dc 100644
 | 
						|
--- a/arch/arm/kernel/entry-header.S
 | 
						|
+++ b/arch/arm/kernel/entry-header.S
 | 
						|
@@ -75,7 +75,11 @@
 | 
						|
 
 | 
						|
 #ifndef CONFIG_THUMB2_KERNEL
 | 
						|
 	.macro	svc_exit, rpsr
 | 
						|
-	msr	spsr_cxsf, \rpsr
 | 
						|
+	cpsid   i
 | 
						|
+	mov     r0, \rpsr
 | 
						|
+	bl      vmm_restore_virq_asm            @ restore the IRQ to emulate
 | 
						|
+	                                        @ the behavior of ldmia {}^
 | 
						|
+	msr	spsr_cxsf, r0
 | 
						|
 #if defined(CONFIG_CPU_V6)
 | 
						|
 	ldr	r0, [sp]
 | 
						|
 	strex	r1, r2, [sp]			@ clear the exclusive monitor
 | 
						|
@@ -90,6 +94,10 @@
 | 
						|
 
 | 
						|
 	.macro	restore_user_regs, fast = 0, offset = 0
 | 
						|
 	ldr	r1, [sp, #\offset + S_PSR]	@ get calling cpsr
 | 
						|
+	@ protect the spsr *and* stack we push the registers into this stack
 | 
						|
+	@ and if the sp is not point to the bottom of the stack, IRQ should be
 | 
						|
+	@ disabled.
 | 
						|
+	cpsid   i
 | 
						|
 	ldr	lr, [sp, #\offset + S_PC]!	@ get pc
 | 
						|
 	msr	spsr_cxsf, r1			@ save in spsr_svc
 | 
						|
 #if defined(CONFIG_CPU_V6)
 | 
						|
@@ -105,6 +113,11 @@
 | 
						|
 	mov	r0, r0				@ ARMv5T and earlier require a nop
 | 
						|
 						@ after ldm {}^
 | 
						|
 	add	sp, sp, #S_FRAME_SIZE - S_PC
 | 
						|
+	@ TODO: in some conditions the call to vmm_on_ret_to_usr is useless.
 | 
						|
+	stmdb   sp!, {r0-r3, ip, lr}
 | 
						|
+	mrs     r0, spsr                        @ debug code
 | 
						|
+	bl      vmm_on_ret_to_usr
 | 
						|
+	ldmia	sp!, {r0-r3, ip, lr}
 | 
						|
 	movs	pc, lr				@ return & move spsr_svc into cpsr
 | 
						|
 	.endm
 | 
						|
 
 | 
						|
diff --git a/arch/arm/mach-omap2/irq.c b/arch/arm/mach-omap2/irq.c
 | 
						|
index 3926f37..252577f 100644
 | 
						|
--- a/arch/arm/mach-omap2/irq.c
 | 
						|
+++ b/arch/arm/mach-omap2/irq.c
 | 
						|
@@ -23,6 +23,10 @@
 | 
						|
 #include <linux/of_address.h>
 | 
						|
 #include <linux/of_irq.h>
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 #include "soc.h"
 | 
						|
 #include "iomap.h"
 | 
						|
 #include "common.h"
 | 
						|
@@ -223,6 +227,14 @@ static inline void omap_intc_handle_irq(void __iomem *base_addr, struct pt_regs
 | 
						|
 {
 | 
						|
 	u32 irqnr;
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+	if (vmm_get_status())
 | 
						|
+	{
 | 
						|
+		vmm_irq_handle(base_addr, domain, regs);
 | 
						|
+		return;
 | 
						|
+	}
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 	do {
 | 
						|
 		irqnr = readl_relaxed(base_addr + 0x98);
 | 
						|
 		if (irqnr)
 | 
						|
diff --git a/arch/arm/mm/fault.c b/arch/arm/mm/fault.c
 | 
						|
index 5dbf13f..e76ba74 100644
 | 
						|
--- a/arch/arm/mm/fault.c
 | 
						|
+++ b/arch/arm/mm/fault.c
 | 
						|
@@ -255,6 +255,10 @@ out:
 | 
						|
 	return fault;
 | 
						|
 }
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 static int __kprobes
 | 
						|
 do_page_fault(unsigned long addr, unsigned int fsr, struct pt_regs *regs)
 | 
						|
 {
 | 
						|
@@ -268,6 +272,11 @@ do_page_fault(unsigned long addr, unsigned int fsr, struct pt_regs *regs)
 | 
						|
 	if (notify_page_fault(regs, fsr))
 | 
						|
 		return 0;
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMMX
 | 
						|
+	WARN(HOST_VMM_ADDR_BEGIN < regs->ARM_pc &&
 | 
						|
+	       regs->ARM_pc < HOST_VMM_ADDR_END);
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 	tsk = current;
 | 
						|
 	mm  = tsk->mm;
 | 
						|
 
 | 
						|
diff --git a/arch/arm/mm/init.c b/arch/arm/mm/init.c
 | 
						|
index ad722f1..ebb4e7f 100644
 | 
						|
--- a/arch/arm/mm/init.c
 | 
						|
+++ b/arch/arm/mm/init.c
 | 
						|
@@ -34,6 +34,10 @@
 | 
						|
 #include <asm/mach/arch.h>
 | 
						|
 #include <asm/mach/map.h>
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 #include "mm.h"
 | 
						|
 
 | 
						|
 static unsigned long phys_initrd_start __initdata = 0;
 | 
						|
@@ -338,6 +342,10 @@ void __init arm_memblock_init(struct meminfo *mi, struct machine_desc *mdesc)
 | 
						|
 	for (i = 0; i < mi->nr_banks; i++)
 | 
						|
 		memblock_add(mi->bank[i].start, mi->bank[i].size);
 | 
						|
 
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+	memblock_reserve(__pa(HOST_VMM_ADDR_BEGIN), HOST_VMM_ADDR_END - HOST_VMM_ADDR_BEGIN);
 | 
						|
+#endif
 | 
						|
+
 | 
						|
 	/* Register the kernel text, kernel data and initrd with memblock. */
 | 
						|
 #ifdef CONFIG_XIP_KERNEL
 | 
						|
 	memblock_reserve(__pa(_sdata), _end - _sdata);
 | 
						|
diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c
 | 
						|
index ce328c7..7e7d0ca 100644
 | 
						|
--- a/arch/arm/mm/mmu.c
 | 
						|
+++ b/arch/arm/mm/mmu.c
 | 
						|
@@ -294,6 +294,20 @@ static struct mem_type mem_types[] = {
 | 
						|
 		.prot_l1   = PMD_TYPE_TABLE,
 | 
						|
 		.domain    = DOMAIN_KERNEL,
 | 
						|
 	},
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+	[MT_RTVMM] = {
 | 
						|
+		.prot_pte  = L_PTE_PRESENT | L_PTE_YOUNG | L_PTE_DIRTY,
 | 
						|
+		.prot_l1   = PMD_TYPE_TABLE,
 | 
						|
+		.prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE,
 | 
						|
+		.domain    = DOMAIN_RTVMM,
 | 
						|
+	},
 | 
						|
+	[MT_RTVMM_SHARE] = {
 | 
						|
+		.prot_pte  = L_PTE_PRESENT | L_PTE_YOUNG | L_PTE_DIRTY,
 | 
						|
+		.prot_l1   = PMD_TYPE_TABLE,
 | 
						|
+		.prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE,
 | 
						|
+		.domain    = DOMAIN_RTVMM_SHR,
 | 
						|
+	},
 | 
						|
+#endif
 | 
						|
 };
 | 
						|
 
 | 
						|
 const struct mem_type *get_mem_type(unsigned int type)
 | 
						|
@@ -450,6 +464,9 @@ static void __init build_mem_type_table(void)
 | 
						|
 			mem_types[MT_DEVICE_CACHED].prot_pte |= L_PTE_SHARED;
 | 
						|
 			mem_types[MT_MEMORY].prot_sect |= PMD_SECT_S;
 | 
						|
 			mem_types[MT_MEMORY].prot_pte |= L_PTE_SHARED;
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+			/* FIXME */
 | 
						|
+#endif
 | 
						|
 			mem_types[MT_MEMORY_DMA_READY].prot_pte |= L_PTE_SHARED;
 | 
						|
 			mem_types[MT_MEMORY_NONCACHED].prot_sect |= PMD_SECT_S;
 | 
						|
 			mem_types[MT_MEMORY_NONCACHED].prot_pte |= L_PTE_SHARED;
 | 
						|
@@ -503,6 +520,12 @@ static void __init build_mem_type_table(void)
 | 
						|
 	mem_types[MT_HIGH_VECTORS].prot_l1 |= ecc_mask;
 | 
						|
 	mem_types[MT_MEMORY].prot_sect |= ecc_mask | cp->pmd;
 | 
						|
 	mem_types[MT_MEMORY].prot_pte |= kern_pgprot;
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+	mem_types[MT_RTVMM].prot_sect |= ecc_mask | cp->pmd;
 | 
						|
+	mem_types[MT_RTVMM].prot_pte |= kern_pgprot;
 | 
						|
+	mem_types[MT_RTVMM_SHARE].prot_sect |= ecc_mask | cp->pmd;
 | 
						|
+	mem_types[MT_RTVMM_SHARE].prot_pte |= kern_pgprot;
 | 
						|
+#endif
 | 
						|
 	mem_types[MT_MEMORY_DMA_READY].prot_pte |= kern_pgprot;
 | 
						|
 	mem_types[MT_MEMORY_NONCACHED].prot_sect |= ecc_mask;
 | 
						|
 	mem_types[MT_ROM].prot_sect |= cp->pmd;
 | 
						|
@@ -1152,6 +1175,27 @@ static void __init devicemaps_init(struct machine_desc *mdesc)
 | 
						|
 #endif
 | 
						|
 
 | 
						|
 	/*
 | 
						|
+	 * Create mappings for RT-Thread VMM and it's shared memory with Linux
 | 
						|
+	 */
 | 
						|
+#ifdef CONFIG_ARM_VMM
 | 
						|
+	/* the TEXCB attribute is not right yet */
 | 
						|
+	/* shared memory region comes first */
 | 
						|
+	map.pfn = __phys_to_pfn(virt_to_phys((void*)HOST_VMM_ADDR_BEGIN));
 | 
						|
+	map.virtual = HOST_VMM_ADDR_BEGIN;
 | 
						|
+	map.length = CONFIG_RTVMM_SHARED_SIZE;
 | 
						|
+	map.type = MT_RTVMM_SHARE;
 | 
						|
+	create_mapping(&map);
 | 
						|
+
 | 
						|
+	/* vmm private region comes next */
 | 
						|
+	map.pfn = __phys_to_pfn(virt_to_phys((void*)HOST_VMM_ADDR_BEGIN
 | 
						|
+					     + CONFIG_RTVMM_SHARED_SIZE));
 | 
						|
+	map.virtual = HOST_VMM_ADDR_BEGIN + CONFIG_RTVMM_SHARED_SIZE;
 | 
						|
+	map.length = CONFIG_HOST_VMM_SIZE - CONFIG_RTVMM_SHARED_SIZE;
 | 
						|
+	map.type = MT_RTVMM;
 | 
						|
+	create_mapping(&map);
 | 
						|
+#endif
 | 
						|
+
 | 
						|
+	/*
 | 
						|
 	 * Create a mapping for the machine vectors at the high-vectors
 | 
						|
 	 * location (0xffff0000).  If we aren't using high-vectors, also
 | 
						|
 	 * create a mapping at the low-vectors virtual address.
 | 
						|
diff --git a/arch/arm/vmm/Kconfig b/arch/arm/vmm/Kconfig
 | 
						|
new file mode 100644
 | 
						|
index 0000000..d852056
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/Kconfig
 | 
						|
@@ -0,0 +1,49 @@
 | 
						|
+menu "RT-Thread VMM Features"
 | 
						|
+
 | 
						|
+# ARM-VMM
 | 
						|
+config ARM_VMM
 | 
						|
+	bool "Support RT-Thread VMM on ARM Cortex-A8"
 | 
						|
+	depends on MACH_REALVIEW_PBA8
 | 
						|
+	help
 | 
						|
+	  RT-Thread VMM implementation on ARM Cortex-A8
 | 
						|
+
 | 
						|
+	  Say Y if you want support for the RT-Thread VMM.
 | 
						|
+	  Otherwise, say N.
 | 
						|
+
 | 
						|
+if SOC_AM33XX
 | 
						|
+config HOST_VMM_ADDR_END
 | 
						|
+	hex "End address of VMM"
 | 
						|
+	depends on ARM_VMM
 | 
						|
+	default 0xE0000000
 | 
						|
+	help
 | 
						|
+	  The end address of VMM space. Normally, it's the
 | 
						|
+	  end address of DDR memory.
 | 
						|
+endif
 | 
						|
+
 | 
						|
+if MACH_REALVIEW_PBA8
 | 
						|
+config HOST_VMM_ADDR_END
 | 
						|
+	hex "End address of VMM"
 | 
						|
+	depends on ARM_VMM
 | 
						|
+	default 0xE0000000
 | 
						|
+	help
 | 
						|
+	  The end address of VMM space. Normally, it's the
 | 
						|
+	  end address of DDR memory.
 | 
						|
+endif
 | 
						|
+
 | 
						|
+config HOST_VMM_SIZE
 | 
						|
+	hex "Size of VMM space"
 | 
						|
+	depends on ARM_VMM
 | 
						|
+	default 0x400000
 | 
						|
+	help
 | 
						|
+	  The size of VMM space.
 | 
						|
+
 | 
						|
+config RTVMM_SHARED_SIZE
 | 
						|
+	hex "Size of shared memory space between rt-vmm and Linux"
 | 
						|
+	depends on ARM_VMM
 | 
						|
+	default 0x100000
 | 
						|
+	help
 | 
						|
+	  The size of shared memory space between rt-vmm and Linux. This shared
 | 
						|
+	  space is within the total size of the HOST_VMM_SIZE. So it is should
 | 
						|
+	  be smaller than HOST_VMM_SIZE.
 | 
						|
+
 | 
						|
+endmenu
 | 
						|
diff --git a/arch/arm/vmm/Makefile b/arch/arm/vmm/Makefile
 | 
						|
new file mode 100644
 | 
						|
index 0000000..127e43a
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/Makefile
 | 
						|
@@ -0,0 +1,10 @@
 | 
						|
+#
 | 
						|
+# Makefile for the linux arm-vmm
 | 
						|
+#
 | 
						|
+
 | 
						|
+obj-$(CONFIG_ARM_VMM) += vmm.o vmm_traps.o vmm_virq.o
 | 
						|
+
 | 
						|
+ifeq ($(CONFIG_ARM_VMM),y)
 | 
						|
+obj-$(CONFIG_SOC_AM33XX)		 += am33xx/softirq.o am33xx/virq.o
 | 
						|
+obj-$(CONFIG_MACH_REALVIEW_PBA8) += realview_a8/softirq.o
 | 
						|
+endif
 | 
						|
diff --git a/arch/arm/vmm/README b/arch/arm/vmm/README
 | 
						|
new file mode 100644
 | 
						|
index 0000000..24f1b42
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/README
 | 
						|
@@ -0,0 +1 @@
 | 
						|
+Linux VMM kernel routines
 | 
						|
diff --git a/arch/arm/vmm/am33xx/intc.h b/arch/arm/vmm/am33xx/intc.h
 | 
						|
new file mode 100644
 | 
						|
index 0000000..6c24f8d
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/am33xx/intc.h
 | 
						|
@@ -0,0 +1,13 @@
 | 
						|
+#ifndef __INTC_H__
 | 
						|
+#define __INTC_H__
 | 
						|
+
 | 
						|
+#define OMAP34XX_IC_BASE	0x48200000
 | 
						|
+
 | 
						|
+#define INTC_SIR_SET0		0x0090
 | 
						|
+#define INTC_MIR_CLEAR0		0x0088
 | 
						|
+
 | 
						|
+#define OMAP2_L4_IO_OFFSET	0xb2000000
 | 
						|
+#define OMAP2_L4_IO_ADDRESS(pa)	IOMEM((pa) + OMAP2_L4_IO_OFFSET) /* L4 */
 | 
						|
+#define OMAP3_IRQ_BASE		OMAP2_L4_IO_ADDRESS(OMAP34XX_IC_BASE)
 | 
						|
+
 | 
						|
+#endif
 | 
						|
diff --git a/arch/arm/vmm/am33xx/softirq.c b/arch/arm/vmm/am33xx/softirq.c
 | 
						|
new file mode 100644
 | 
						|
index 0000000..5648496
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/am33xx/softirq.c
 | 
						|
@@ -0,0 +1,14 @@
 | 
						|
+#include <linux/kernel.h>
 | 
						|
+#include <linux/module.h>
 | 
						|
+#include <asm/io.h>
 | 
						|
+
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+#include "../vmm_virhw.h"
 | 
						|
+#include "intc.h"
 | 
						|
+
 | 
						|
+void vmm_raise_softirq(int irq)
 | 
						|
+{
 | 
						|
+	writel_relaxed(1 << (irq % 32),
 | 
						|
+		OMAP3_IRQ_BASE + INTC_SIR_SET0 + (irq / 32) * 4);
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_raise_softirq);
 | 
						|
diff --git a/arch/arm/vmm/am33xx/virq.c b/arch/arm/vmm/am33xx/virq.c
 | 
						|
new file mode 100644
 | 
						|
index 0000000..4ef7671
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/am33xx/virq.c
 | 
						|
@@ -0,0 +1,48 @@
 | 
						|
+#include <linux/kernel.h>
 | 
						|
+#include <linux/module.h>
 | 
						|
+#include <linux/irqdomain.h>
 | 
						|
+
 | 
						|
+#include <asm/io.h>
 | 
						|
+#include <asm/irq.h>
 | 
						|
+
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+#include "../vmm_virhw.h"
 | 
						|
+#include "intc.h"
 | 
						|
+
 | 
						|
+void vmm_irq_handle(void __iomem *base_addr, struct irq_domain *domain,
 | 
						|
+				 struct pt_regs *regs)
 | 
						|
+{
 | 
						|
+	unsigned long flags;
 | 
						|
+	struct vmm_context* _vmm_context;
 | 
						|
+
 | 
						|
+	_vmm_context = vmm_context_get();
 | 
						|
+
 | 
						|
+	while (_vmm_context->virq_pended) {
 | 
						|
+		int index;
 | 
						|
+
 | 
						|
+		flags = vmm_irq_save();
 | 
						|
+		_vmm_context->virq_pended = 0;
 | 
						|
+		vmm_irq_restore(flags);
 | 
						|
+
 | 
						|
+		/* get the pending interrupt */
 | 
						|
+		for (index = 0; index < IRQS_NR_32; index++) {
 | 
						|
+			int pdbit;
 | 
						|
+
 | 
						|
+			for (pdbit = __builtin_ffs(_vmm_context->virq_pending[index]);
 | 
						|
+			     pdbit != 0;
 | 
						|
+			     pdbit = __builtin_ffs(_vmm_context->virq_pending[index])) {
 | 
						|
+				unsigned long inner_flag;
 | 
						|
+				int irqnr;
 | 
						|
+
 | 
						|
+				pdbit--;
 | 
						|
+
 | 
						|
+				inner_flag = vmm_irq_save();
 | 
						|
+				_vmm_context->virq_pending[index] &= ~(1 << pdbit);
 | 
						|
+				vmm_irq_restore(inner_flag);
 | 
						|
+
 | 
						|
+				irqnr = irq_find_mapping(domain, pdbit + index * 32);
 | 
						|
+				handle_IRQ(irqnr, regs);
 | 
						|
+			}
 | 
						|
+		}
 | 
						|
+	}
 | 
						|
+}
 | 
						|
diff --git a/arch/arm/vmm/realview_a8/softirq.c b/arch/arm/vmm/realview_a8/softirq.c
 | 
						|
new file mode 100644
 | 
						|
index 0000000..a52b79c7
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/realview_a8/softirq.c
 | 
						|
@@ -0,0 +1,12 @@
 | 
						|
+#include <linux/kernel.h>
 | 
						|
+#include <linux/module.h>
 | 
						|
+#include <asm/io.h>
 | 
						|
+#include <asm/hardware/gic.h>
 | 
						|
+
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+
 | 
						|
+void vmm_raise_softirq(int irq)
 | 
						|
+{
 | 
						|
+	gic_raise_softirq(cpumask_of(0),  irq);
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_raise_softirq);
 | 
						|
diff --git a/arch/arm/vmm/vmm.c b/arch/arm/vmm/vmm.c
 | 
						|
new file mode 100644
 | 
						|
index 0000000..3b1d202
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/vmm.c
 | 
						|
@@ -0,0 +1,32 @@
 | 
						|
+#include <linux/kernel.h>
 | 
						|
+#include <linux/module.h>
 | 
						|
+
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+
 | 
						|
+struct vmm_context* _vmm_context = NULL;
 | 
						|
+int vmm_status = 0;
 | 
						|
+EXPORT_SYMBOL(vmm_status);
 | 
						|
+
 | 
						|
+void vmm_set_status(int status)
 | 
						|
+{
 | 
						|
+	vmm_status = status;
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_set_status);
 | 
						|
+
 | 
						|
+int vmm_get_status(void)
 | 
						|
+{
 | 
						|
+	return vmm_status;
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_get_status);
 | 
						|
+
 | 
						|
+void vmm_context_init(void* context_addr)
 | 
						|
+{
 | 
						|
+	_vmm_context = (struct vmm_context*)context_addr;
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_context_init);
 | 
						|
+
 | 
						|
+struct vmm_context* vmm_context_get(void)
 | 
						|
+{
 | 
						|
+	return _vmm_context;
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_context_get);
 | 
						|
diff --git a/arch/arm/vmm/vmm_traps.c b/arch/arm/vmm/vmm_traps.c
 | 
						|
new file mode 100644
 | 
						|
index 0000000..def0d90
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/vmm_traps.c
 | 
						|
@@ -0,0 +1,37 @@
 | 
						|
+#include <linux/kernel.h>
 | 
						|
+#include <linux/module.h>
 | 
						|
+#include <asm/traps.h>
 | 
						|
+#include <asm/cp15.h>
 | 
						|
+#include <asm/cacheflush.h>
 | 
						|
+
 | 
						|
+void trap_set_vector(void *start, unsigned int length)
 | 
						|
+{
 | 
						|
+	unsigned char *ptr;
 | 
						|
+	unsigned char *vector;
 | 
						|
+
 | 
						|
+	ptr = start;
 | 
						|
+	vector = (unsigned char*)vectors_page;
 | 
						|
+
 | 
						|
+	/* only set IRQ and FIQ */
 | 
						|
+#if defined(CONFIG_CPU_USE_DOMAINS)
 | 
						|
+	/* IRQ */
 | 
						|
+	memcpy((void *)0xffff0018, (void*)(ptr + 0x18), 4);
 | 
						|
+	memcpy((void *)(0xffff0018 + 0x20), (void*)(ptr + 0x18 + 0x20), 4);
 | 
						|
+
 | 
						|
+	/* FIQ */
 | 
						|
+	memcpy((void *)0xffff001C, (void*)(ptr + 0x1C), 4);
 | 
						|
+	memcpy((void *)(0xffff001C + 0x20), (void*)(ptr + 0x1C + 0x20), 4);
 | 
						|
+#else
 | 
						|
+	/* IRQ */
 | 
						|
+	memcpy(vector + 0x18, (void*)(ptr + 0x18), 4);
 | 
						|
+	memcpy(vector + 0x18 + 0x20, (void*)(ptr + 0x18 + 0x20), 4);
 | 
						|
+
 | 
						|
+	/* FIQ */
 | 
						|
+	memcpy(vector + 0x1C, (void*)(ptr + 0x1C), 4);
 | 
						|
+	memcpy(vector + 0x1C + 0x20, (void*)(ptr + 0x1C + 0x20), 4);
 | 
						|
+#endif
 | 
						|
+	flush_icache_range(0xffff0000, 0xffff0000 + length);
 | 
						|
+	if (!vectors_high())
 | 
						|
+		flush_icache_range(0x00, 0x00 + length);
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(trap_set_vector);
 | 
						|
diff --git a/arch/arm/vmm/vmm_virhw.h b/arch/arm/vmm/vmm_virhw.h
 | 
						|
new file mode 100644
 | 
						|
index 0000000..363cc6e
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/vmm_virhw.h
 | 
						|
@@ -0,0 +1,59 @@
 | 
						|
+#ifndef __VMM_VIRTHWH__
 | 
						|
+#define __VMM_VIRTHWH__
 | 
						|
+
 | 
						|
+#define REALVIEW_NR_IRQS        96
 | 
						|
+#define IRQS_NR_32              ((REALVIEW_NR_IRQS + 31)/32)
 | 
						|
+#define RTT_VMM_IRQ_TRIGGER     10
 | 
						|
+
 | 
						|
+struct vmm_context
 | 
						|
+{
 | 
						|
+	/* the status of vGuest irq */
 | 
						|
+	volatile unsigned long virq_status;
 | 
						|
+
 | 
						|
+	/* has interrupt pended on vGuest OS IRQ */
 | 
						|
+	volatile unsigned long virq_pended;
 | 
						|
+
 | 
						|
+	/* pending interrupt for vGuest OS */
 | 
						|
+	volatile unsigned long virq_pending[IRQS_NR_32];
 | 
						|
+};
 | 
						|
+
 | 
						|
+/* IRQ operation under VMM */
 | 
						|
+static inline unsigned long vmm_irq_save(void)
 | 
						|
+{
 | 
						|
+	unsigned long flags;
 | 
						|
+
 | 
						|
+	asm volatile(
 | 
						|
+		"	mrs	%0, cpsr	@ arch_local_irq_save\n"
 | 
						|
+		"	cpsid	i"
 | 
						|
+		: "=r" (flags) : : "memory", "cc");
 | 
						|
+	return flags;
 | 
						|
+}
 | 
						|
+
 | 
						|
+static inline void vmm_irq_restore(unsigned long flags)
 | 
						|
+{
 | 
						|
+	asm volatile(
 | 
						|
+		"	msr	cpsr_c, %0	@ local_irq_restore"
 | 
						|
+		:
 | 
						|
+		: "r" (flags)
 | 
						|
+		: "memory", "cc");
 | 
						|
+}
 | 
						|
+
 | 
						|
+static inline void vmm_irq_enable(void)
 | 
						|
+{
 | 
						|
+	asm volatile(
 | 
						|
+		"	cpsie i			@ arch_local_irq_enable"
 | 
						|
+		:
 | 
						|
+		:
 | 
						|
+		: "memory", "cc");
 | 
						|
+}
 | 
						|
+
 | 
						|
+static inline void vmm_irq_disable(void)
 | 
						|
+{
 | 
						|
+	asm volatile(
 | 
						|
+		"	cpsid i			@ arch_local_irq_disable"
 | 
						|
+		:
 | 
						|
+		:
 | 
						|
+		: "memory", "cc");
 | 
						|
+}
 | 
						|
+
 | 
						|
+#endif
 | 
						|
diff --git a/arch/arm/vmm/vmm_virq.c b/arch/arm/vmm/vmm_virq.c
 | 
						|
new file mode 100644
 | 
						|
index 0000000..85886a2
 | 
						|
--- /dev/null
 | 
						|
+++ b/arch/arm/vmm/vmm_virq.c
 | 
						|
@@ -0,0 +1,183 @@
 | 
						|
+#include <linux/bug.h>
 | 
						|
+#include <linux/kernel.h>
 | 
						|
+#include <linux/module.h>
 | 
						|
+#include <asm/unwind.h>
 | 
						|
+
 | 
						|
+#include <vmm/vmm.h>
 | 
						|
+
 | 
						|
+#include "vmm_virhw.h"
 | 
						|
+
 | 
						|
+/* VMM use the I bit in SPSR to save the virq status in the isr entry. So warn
 | 
						|
+ * on the I bit set would gave some false negative result. */
 | 
						|
+//#define VMM_WARN_ON_I_BIT
 | 
						|
+
 | 
						|
+extern struct vmm_context* _vmm_context;
 | 
						|
+
 | 
						|
+void vmm_disable_virq(void)
 | 
						|
+{
 | 
						|
+	unsigned long flags = vmm_irq_save();
 | 
						|
+	_vmm_context->virq_status = 0x01;
 | 
						|
+	vmm_irq_restore(flags);
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_disable_virq);
 | 
						|
+
 | 
						|
+static void _vmm_raise_on_pended(void)
 | 
						|
+{
 | 
						|
+	/* check any interrupt pended in vIRQ */
 | 
						|
+	if (_vmm_context->virq_pended) {
 | 
						|
+		/* trigger an soft interrupt */
 | 
						|
+		vmm_raise_softirq(RTT_VMM_IRQ_TRIGGER);
 | 
						|
+		return;
 | 
						|
+	}
 | 
						|
+
 | 
						|
+#if 0
 | 
						|
+	int i;
 | 
						|
+	for (i = 0; i < ARRAY_SIZE(_vmm_context->virq_pending); i++) {
 | 
						|
+		if (_vmm_context->virq_pending[i]) {
 | 
						|
+			_vmm_context->virq_pended = 1;
 | 
						|
+			pr_info("\n");
 | 
						|
+			vmm_raise_softirq(RTT_VMM_IRQ_TRIGGER);
 | 
						|
+			return;
 | 
						|
+		}
 | 
						|
+	}
 | 
						|
+#endif
 | 
						|
+}
 | 
						|
+
 | 
						|
+void vmm_enable_virq(void)
 | 
						|
+{
 | 
						|
+	unsigned long flags = vmm_irq_save();
 | 
						|
+	_vmm_context->virq_status = 0x00;
 | 
						|
+	_vmm_raise_on_pended();
 | 
						|
+	vmm_irq_restore(flags);
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_enable_virq);
 | 
						|
+
 | 
						|
+unsigned long vmm_return_virq(void)
 | 
						|
+{
 | 
						|
+	unsigned long flags;
 | 
						|
+	unsigned long level;
 | 
						|
+
 | 
						|
+	level =  vmm_irq_save();
 | 
						|
+	flags = _vmm_context->virq_status;
 | 
						|
+	vmm_irq_restore(level);
 | 
						|
+
 | 
						|
+	return flags;
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_return_virq);
 | 
						|
+
 | 
						|
+unsigned long vmm_save_virq(void)
 | 
						|
+{
 | 
						|
+	int status;
 | 
						|
+	unsigned long flags = vmm_irq_save();
 | 
						|
+
 | 
						|
+	status = _vmm_context->virq_status;
 | 
						|
+	_vmm_context->virq_status = 0x01;
 | 
						|
+	vmm_irq_restore(flags);
 | 
						|
+
 | 
						|
+	return status;
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_save_virq);
 | 
						|
+
 | 
						|
+void vmm_restore_virq(unsigned long flags)
 | 
						|
+{
 | 
						|
+	unsigned long level;
 | 
						|
+
 | 
						|
+	level =  vmm_irq_save();
 | 
						|
+	_vmm_context->virq_status = flags;
 | 
						|
+	if (_vmm_context->virq_status == 0)
 | 
						|
+	{
 | 
						|
+		_vmm_raise_on_pended();
 | 
						|
+	}
 | 
						|
+	vmm_irq_restore(level);
 | 
						|
+}
 | 
						|
+EXPORT_SYMBOL(vmm_restore_virq);
 | 
						|
+
 | 
						|
+unsigned long vmm_save_virq_spsr_asm(unsigned long spsr, struct pt_regs *regs)
 | 
						|
+{
 | 
						|
+	if (vmm_status) {
 | 
						|
+		if (_vmm_context->virq_status)
 | 
						|
+			return spsr | PSR_I_BIT;
 | 
						|
+	}
 | 
						|
+	return spsr;
 | 
						|
+}
 | 
						|
+
 | 
						|
+void irq_enable_asm(void)
 | 
						|
+{
 | 
						|
+	if (vmm_status) {
 | 
						|
+		vmm_enable_virq();
 | 
						|
+	} else {
 | 
						|
+		asm volatile("cpsie i" : : : "memory", "cc");
 | 
						|
+	}
 | 
						|
+}
 | 
						|
+
 | 
						|
+void irq_disable_asm(void)
 | 
						|
+{
 | 
						|
+	if (vmm_status) {
 | 
						|
+		vmm_disable_virq();
 | 
						|
+	} else {
 | 
						|
+		asm volatile("cpsid i" : : : "memory", "cc");
 | 
						|
+	}
 | 
						|
+}
 | 
						|
+
 | 
						|
+/* should be called when the guest entering the state that the IRQ is disabled
 | 
						|
+ * by hardware, for example, entering SVC, PABT, DABT mode.
 | 
						|
+ *
 | 
						|
+ * It will the open the hardware IRQ, virtual IRQ remain unchanged.
 | 
						|
+ */
 | 
						|
+void vmm_switch_nohwirq_to_novirq(void)
 | 
						|
+{
 | 
						|
+	if (vmm_status) {
 | 
						|
+		vmm_disable_virq();
 | 
						|
+		asm volatile("cpsie i" : : : "memory", "cc");
 | 
						|
+	}
 | 
						|
+}
 | 
						|
+
 | 
						|
+unsigned long vmm_restore_virq_asm(unsigned long spsr)
 | 
						|
+{
 | 
						|
+	if (vmm_status) {
 | 
						|
+#ifdef VMM_WARN_ON_I_BIT
 | 
						|
+		WARN(spsr & PSR_I_BIT, "return to svc mode with I in SPSR set\n");
 | 
						|
+#endif
 | 
						|
+		vmm_restore_virq(!!(spsr & PSR_I_BIT));
 | 
						|
+		return spsr & ~PSR_I_BIT;
 | 
						|
+	} else {
 | 
						|
+		return spsr;
 | 
						|
+	}
 | 
						|
+}
 | 
						|
+
 | 
						|
+void vmm_on_ret_to_usr(unsigned long spsr)
 | 
						|
+{
 | 
						|
+	if (vmm_status) {
 | 
						|
+#ifdef VMM_WARN_ON_I_BIT
 | 
						|
+		WARN(spsr & PSR_I_BIT, "return to user mode with I in SPSR set\n");
 | 
						|
+#endif
 | 
						|
+		vmm_enable_virq();
 | 
						|
+	}
 | 
						|
+}
 | 
						|
+
 | 
						|
+void vmm_on_svc_exit_irq(unsigned long spsr)
 | 
						|
+{
 | 
						|
+	if (vmm_status) {
 | 
						|
+#ifdef VMM_WARN_ON_I_BIT
 | 
						|
+		WARN(spsr & PSR_I_BIT, "exit IRQ with I in SPSR set\n");
 | 
						|
+#endif
 | 
						|
+		vmm_enable_virq();
 | 
						|
+	}
 | 
						|
+}
 | 
						|
+
 | 
						|
+void vmm_dump_irq(void)
 | 
						|
+{
 | 
						|
+	int i;
 | 
						|
+	unsigned long cpsr;
 | 
						|
+
 | 
						|
+	asm volatile ("mrs %0, cpsr": "=r"(cpsr));
 | 
						|
+
 | 
						|
+	printk("status: %08lx, pended: %08lx, cpsr: %08lx\n",
 | 
						|
+	       _vmm_context->virq_status, _vmm_context->virq_pended, cpsr);
 | 
						|
+	printk("pending: ");
 | 
						|
+	for (i = 0; i < ARRAY_SIZE(_vmm_context->virq_pending); i++) {
 | 
						|
+		printk("%08lx, ", _vmm_context->virq_pending[i]);
 | 
						|
+	}
 | 
						|
+	printk("\n");
 | 
						|
+}
 | 
						|
+
 | 
						|
-- 
 | 
						|
1.8.4
 | 
						|
 |