/* * Machine vector for IA-64. * * Copyright (C) 1999 Silicon Graphics, Inc. * Copyright (C) Srinivasa Thirumalachar * Copyright (C) Vijay Chander * Copyright (C) 1999-2001, 2003-2004 Hewlett-Packard Co. * David Mosberger-Tang */ #ifndef _ASM_IA64_MACHVEC_H #define _ASM_IA64_MACHVEC_H #include #include /* forward declarations: */ struct device; struct pt_regs; struct scatterlist; struct page; struct mm_struct; struct pci_bus; struct task_struct; struct pci_dev; struct msi_desc; struct dma_attrs; typedef void ia64_mv_setup_t (char **); typedef void ia64_mv_cpu_init_t (void); typedef void ia64_mv_irq_init_t (void); typedef void ia64_mv_send_ipi_t (int, int, int, int); typedef void ia64_mv_timer_interrupt_t (int, void *); typedef void ia64_mv_global_tlb_purge_t (struct mm_struct *, unsigned long, unsigned long, unsigned long); typedef void ia64_mv_tlb_migrate_finish_t (struct mm_struct *); typedef u8 ia64_mv_irq_to_vector (int); typedef unsigned int ia64_mv_local_vector_to_irq (u8); typedef char *ia64_mv_pci_get_legacy_mem_t (struct pci_bus *); typedef int ia64_mv_pci_legacy_read_t (struct pci_bus *, u16 port, u32 *val, u8 size); typedef int ia64_mv_pci_legacy_write_t (struct pci_bus *, u16 port, u32 val, u8 size); typedef void ia64_mv_migrate_t(struct task_struct * task); typedef void ia64_mv_pci_fixup_bus_t (struct pci_bus *); typedef void ia64_mv_kernel_launch_event_t(void); /* DMA-mapping interface: */ typedef void ia64_mv_dma_init (void); typedef void *ia64_mv_dma_alloc_coherent (struct device *, size_t, dma_addr_t *, gfp_t); typedef void ia64_mv_dma_free_coherent (struct device *, size_t, void *, dma_addr_t); typedef dma_addr_t ia64_mv_dma_map_single (struct device *, void *, size_t, int); typedef void ia64_mv_dma_unmap_single (struct device *, dma_addr_t, size_t, int); typedef int ia64_mv_dma_map_sg (struct device *, struct scatterlist *, int, int); typedef void ia64_mv_dma_unmap_sg (struct device *, struct scatterlist *, int, int); typedef void ia64_mv_dma_sync_single_for_cpu (struct device *, dma_addr_t, size_t, int); typedef void ia64_mv_dma_sync_sg_for_cpu (struct device *, struct scatterlist *, int, int); typedef void ia64_mv_dma_sync_single_for_device (struct device *, dma_addr_t, size_t, int); typedef void ia64_mv_dma_sync_sg_for_device (struct device *, struct scatterlist *, int, int); typedef int ia64_mv_dma_mapping_error(struct device *, dma_addr_t dma_addr); typedef int ia64_mv_dma_supported (struct device *, u64); typedef dma_addr_t ia64_mv_dma_map_single_attrs (struct device *, void *, size_t, int, struct dma_attrs *); typedef void ia64_mv_dma_unmap_single_attrs (struct device *, dma_addr_t, size_t, int, struct dma_attrs *); typedef int ia64_mv_dma_map_sg_attrs (struct device *, struct scatterlist *, int, int, struct dma_attrs *); typedef void ia64_mv_dma_unmap_sg_attrs (struct device *, struct scatterlist *, int, int, struct dma_attrs *); /* * WARNING: The legacy I/O space is _architected_. Platforms are * expected to follow this architected model (see Section 10.7 in the * IA-64 Architecture Software Developer's Manual). Unfortunately, * some broken machines do not follow that model, which is why we have * to make the inX/outX operations part of the machine vector. * Platform designers should follow the architected model whenever * possible. */ typedef unsigned int ia64_mv_inb_t (unsigned long); typedef unsigned int ia64_mv_inw_t (unsigned long); typedef unsigned int ia64_mv_inl_t (unsigned long); typedef void ia64_mv_outb_t (unsigned char, unsigned long); typedef void ia64_mv_outw_t (unsigned short, unsigned long); typedef void ia64_mv_outl_t (unsigned int, unsigned long); typedef void ia64_mv_mmiowb_t (void); typedef unsigned char ia64_mv_readb_t (const volatile void __iomem *); typedef unsigned short ia64_mv_readw_t (const volatile void __iomem *); typedef unsigned int ia64_mv_readl_t (const volatile void __iomem *); typedef unsigned long ia64_mv_readq_t (const volatile void __iomem *); typedef unsigned char ia64_mv_readb_relaxed_t (const volatile void __iomem *); typedef unsigned short ia64_mv_readw_relaxed_t (const volatile void __iomem *); typedef unsigned int ia64_mv_readl_relaxed_t (const volatile void __iomem *); typedef unsigned long ia64_mv_readq_relaxed_t (const volatile void __iomem *); typedef int ia64_mv_setup_msi_irq_t (struct pci_dev *pdev, struct msi_desc *); typedef void ia64_mv_teardown_msi_irq_t (unsigned int irq); static inline void machvec_noop (void) { } static inline void machvec_noop_mm (struct mm_struct *mm) { } static inline void machvec_noop_task (struct task_struct *task) { } static inline void machvec_noop_bus (struct pci_bus *bus) { } extern void machvec_setup (char **); extern void machvec_timer_interrupt (int, void *); extern void machvec_dma_sync_single (struct device *, dma_addr_t, size_t, int); extern void machvec_dma_sync_sg (struct device *, struct scatterlist *, int, int); extern void machvec_tlb_migrate_finish (struct mm_struct *); # if defined (CONFIG_IA64_HP_SIM) # include # elif defined (CONFIG_IA64_DIG) # include # elif defined(CONFIG_IA64_DIG_VTD) # include # elif defined (CONFIG_IA64_HP_ZX1) # include # elif defined (CONFIG_IA64_HP_ZX1_SWIOTLB) # include # elif defined (CONFIG_IA64_SGI_SN2) # include # elif defined (CONFIG_IA64_SGI_UV) # include # elif defined (CONFIG_IA64_XEN_GUEST) # include # elif defined (CONFIG_IA64_GENERIC) # ifdef MACHVEC_PLATFORM_HEADER # include MACHVEC_PLATFORM_HEADER # else # define platform_name ia64_mv.name # define platform_setup ia64_mv.setup # define platform_cpu_init ia64_mv.cpu_init # define platform_irq_init ia64_mv.irq_init # define platform_send_ipi ia64_mv.send_ipi # define platform_timer_interrupt ia64_mv.timer_interrupt # define platform_global_tlb_purge ia64_mv.global_tlb_purge # define platform_tlb_migrate_finish ia64_mv.tlb_migrate_finish # define platform_dma_init ia64_mv.dma_init # define platform_dma_alloc_coherent ia64_mv.dma_alloc_coherent # define platform_dma_free_coherent ia64_mv.dma_free_coherent # define platform_dma_map_single_attrs ia64_mv.dma_map_single_attrs # define platform_dma_unmap_single_attrs ia64_mv.dma_unmap_single_attrs # define platform_dma_map_sg_attrs ia64_mv.dma_map_sg_attrs # define platform_dma_unmap_sg_attrs ia64_mv.dma_unmap_sg_attrs # define platform_dma_sync_single_for_cpu ia64_mv.dma_sync_single_for_cpu # define platform_dma_sync_sg_for_cpu ia64_mv.dma_sync_sg_for_cpu # define platform_dma_sync_single_for_device ia64_mv.dma_sync_single_for_device # define platform_dma_sync_sg_for_device ia64_mv.dma_sync_sg_for_device # define platform_dma_mapping_error ia64_mv.dma_mapping_error # define platform_dma_supported ia64_mv.dma_supported # define platform_irq_to_vector ia64_mv.irq_to_vector # define platform_local_vector_to_irq ia64_mv.local_vector_to_irq # define platform_pci_get_legacy_mem ia64_mv.pci_get_legacy_mem # define platform_pci_legacy_read ia64_mv.pci_legacy_read # define platform_pci_legacy_write ia64_mv.pci_legacy_write # define platform_inb ia64_mv.inb # define platform_inw ia64_mv.inw # define platform_inl ia64_mv.inl # define platform_outb ia64_mv.outb # define platform_outw ia64_mv.outw # define platform_outl ia64_mv.outl # define platform_mmiowb ia64_mv.mmiowb # define platform_readb ia64_mv.readb # define platform_readw ia64_mv.readw # define platform_readl ia64_mv.readl # define platform_readq ia64_mv.readq # define platform_readb_relaxed ia64_mv.readb_relaxed # define platform_readw_relaxed ia64_mv.readw_relaxed # define platform_readl_relaxed ia64_mv.readl_relaxed # define platform_readq_relaxed ia64_mv.readq_relaxed # define platform_migrate ia64_mv.migrate # define platform_setup_msi_irq ia64_mv.setup_msi_irq # define platform_teardown_msi_irq ia64_mv.teardown_msi_irq # define platform_pci_fixup_bus ia64_mv.pci_fixup_bus # define platform_kernel_launch_event ia64_mv.kernel_launch_event # endif /* __attribute__((__aligned__(16))) is required to make size of the * structure multiple of 16 bytes. * This will fillup the holes created because of section 3.3.1 in * Software Conventions guide. */ struct ia64_machine_vector { const char *name; ia64_mv_setup_t *setup; ia64_mv_cpu_init_t *cpu_init; ia64_mv_irq_init_t *irq_init; ia64_mv_send_ipi_t *send_ipi; ia64_mv_timer_interrupt_t *timer_interrupt; ia64_mv_global_tlb_purge_t *global_tlb_purge; ia64_mv_tlb_migrate_finish_t *tlb_migrate_finish; ia64_mv_dma_init *dma_init; ia64_mv_dma_alloc_coherent *dma_alloc_coherent; ia64_mv_dma_free_coherent *dma_free_coherent; ia64_mv_dma_map_single_attrs *dma_map_single_attrs; ia64_mv_dma_unmap_single_attrs *dma_unmap_single_attrs; ia64_mv_dma_map_sg_attrs *dma_map_sg_attrs; ia64_mv_dma_unmap_sg_attrs *dma_unmap_sg_attrs; ia64_mv_dma_sync_single_for_cpu *dma_sync_single_for_cpu; ia64_mv_dma_sync_sg_for_cpu *dma_sync_sg_for_cpu; ia64_mv_dma_sync_single_for_device *dma_sync_single_for_device; ia64_mv_dma_sync_sg_for_device *dma_sync_sg_for_device; ia64_mv_dma_mapping_error *dma_mapping_error; ia64_mv_dma_supported *dma_supported; ia64_mv_irq_to_vector *irq_to_vector; ia64_mv_local_vector_to_irq *local_vector_to_irq; ia64_mv_pci_get_legacy_mem_t *pci_get_legacy_mem; ia64_mv_pci_legacy_read_t *pci_legacy_read; ia64_mv_pci_legacy_write_t *pci_legacy_write; ia64_mv_inb_t *inb; ia64_mv_inw_t *inw; ia64_mv_inl_t *inl; ia64_mv_outb_t *outb; ia64_mv_outw_t *outw; ia64_mv_outl_t *outl; ia64_mv_mmiowb_t *mmiowb; ia64_mv_readb_t *readb; ia64_mv_readw_t *readw; ia64_mv_readl_t *readl; ia64_mv_readq_t *readq; ia64_mv_readb_relaxed_t *readb_relaxed; ia64_mv_readw_relaxed_t *readw_relaxed; ia64_mv_readl_relaxed_t *readl_relaxed; ia64_mv_readq_relaxed_t *readq_relaxed; ia64_mv_migrate_t *migrate; ia64_mv_setup_msi_irq_t *setup_msi_irq; ia64_mv_teardown_msi_irq_t *teardown_msi_irq; ia64_mv_pci_fixup_bus_t *pci_fixup_bus; ia64_mv_kernel_launch_event_t *kernel_launch_event; } __attribute__((__aligned__(16))); /* align attrib? see above comment */ #define MACHVEC_INIT(name) \ { \ #name, \ platform_setup, \ platform_cpu_init, \ platform_irq_init, \ platform_send_ipi, \ platform_timer_interrupt, \ platform_global_tlb_purge, \ platform_tlb_migrate_finish, \ platform_dma_init, \ platform_dma_alloc_coherent, \ platform_dma_free_coherent, \ platform_dma_map_single_attrs, \ platform_dma_unmap_single_attrs, \ platform_dma_map_sg_attrs, \ platform_dma_unmap_sg_attrs, \ platform_dma_sync_single_for_cpu, \ platform_dma_sync_sg_for_cpu, \ platform_dma_sync_single_for_device, \ platform_dma_sync_sg_for_device, \ platform_dma_mapping_error, \ platform_dma_supported, \ platform_irq_to_vector, \ platform_local_vector_to_irq, \ platform_pci_get_legacy_mem, \ platform_pci_legacy_read, \ platform_pci_legacy_write, \ platform_inb, \ platform_inw, \ platform_inl, \ platform_outb, \ platform_outw, \ platform_outl, \ platform_mmiowb, \ platform_readb, \ platform_readw, \ platform_readl, \ platform_readq, \ platform_readb_relaxed, \ platform_readw_relaxed, \ platform_readl_relaxed, \ platform_readq_relaxed, \ platform_migrate, \ platform_setup_msi_irq, \ platform_teardown_msi_irq, \ platform_pci_fixup_bus, \ platform_kernel_launch_event \ } extern struct ia64_machine_vector ia64_mv; extern void machvec_init (const char *name); extern void machvec_init_from_cmdline(const char *cmdline); # else # error Unknown configuration. Update arch/ia64/include/asm/machvec.h. # endif /* CONFIG_IA64_GENERIC */ extern void swiotlb_dma_init(void); /* * Define default versions so we can extend machvec for new platforms without having * to update the machvec files for all existing platforms. */ #ifndef platform_setup # define platform_setup machvec_setup #endif #ifndef platform_cpu_init # define platform_cpu_init machvec_noop #endif #ifndef platform_irq_init # define platform_irq_init machvec_noop #endif #ifndef platform_send_ipi # define platform_send_ipi ia64_send_ipi /* default to architected version */ #endif #ifndef platform_timer_interrupt # define platform_timer_interrupt machvec_timer_interrupt #endif #ifndef platform_global_tlb_purge # define platform_global_tlb_purge ia64_global_tlb_purge /* default to architected version */ #endif #ifndef platform_tlb_migrate_finish # define platform_tlb_migrate_finish machvec_noop_mm #endif #ifndef platform_kernel_launch_event # define platform_kernel_launch_event machvec_noop #endif #ifndef platform_dma_init # define platform_dma_init swiotlb_dma_init #endif #ifndef platform_dma_alloc_coherent # define platform_dma_alloc_coherent swiotlb_alloc_coherent #endif #ifndef platform_dma_free_coherent # define platform_dma_free_coherent swiotlb_free_coherent #endif #ifndef platform_dma_map_single_attrs # define platform_dma_map_single_attrs swiotlb_map_single_attrs #endif #ifndef platform_dma_unmap_single_attrs # define platform_dma_unmap_single_attrs swiotlb_unmap_single_attrs #endif #ifndef platform_dma_map_sg_attrs # define platform_dma_map_sg_attrs swiotlb_map_sg_attrs #endif #ifndef platform_dma_unmap_sg_attrs # define platform_dma_unmap_sg_attrs swiotlb_unmap_sg_attrs #endif #ifndef platform_dma_sync_single_for_cpu # define platform_dma_sync_single_for_cpu swiotlb_sync_single_for_cpu #endif #ifndef platform_dma_sync_sg_for_cpu # define platform_dma_sync_sg_for_cpu swiotlb_sync_sg_for_cpu #endif #ifndef platform_dma_sync_single_for_device # define platform_dma_sync_single_for_device swiotlb_sync_single_for_device #endif #ifndef platform_dma_sync_sg_for_device # define platform_dma_sync_sg_for_device swiotlb_sync_sg_for_device #endif #ifndef platform_dma_mapping_error # define platform_dma_mapping_error swiotlb_dma_mapping_error #endif #ifndef platform_dma_supported # define platform_dma_supported swiotlb_dma_supported #endif #ifndef platform_irq_to_vector # define platform_irq_to_vector __ia64_irq_to_vector #endif #ifndef platform_local_vector_to_irq # define platform_local_vector_to_irq __ia64_local_vector_to_irq #endif #ifndef platform_pci_get_legacy_mem # define platform_pci_get_legacy_mem ia64_pci_get_legacy_mem #endif #ifndef platform_pci_legacy_read # define platform_pci_legacy_read ia64_pci_legacy_read extern int ia64_pci_legacy_read(struct pci_bus *bus, u16 port, u32 *val, u8 size); #endif #ifndef platform_pci_legacy_write # define platform_pci_legacy_write ia64_pci_legacy_write extern int ia64_pci_legacy_write(struct pci_bus *bus, u16 port, u32 val, u8 size); #endif #ifndef platform_inb # define platform_inb __ia64_inb #endif #ifndef platform_inw # define platform_inw __ia64_inw #endif #ifndef platform_inl # define platform_inl __ia64_inl #endif #ifndef platform_outb # define platform_outb __ia64_outb #endif #ifndef platform_outw # define platform_outw __ia64_outw #endif #ifndef platform_outl # define platform_outl __ia64_outl #endif #ifndef platform_mmiowb # define platform_mmiowb __ia64_mmiowb #endif #ifndef platform_readb # define platform_readb __ia64_readb #endif #ifndef platform_readw # define platform_readw __ia64_readw #endif #ifndef platform_readl # define platform_readl __ia64_readl #endif #ifndef platform_readq # define platform_readq __ia64_readq #endif #ifndef platform_readb_relaxed # define platform_readb_relaxed __ia64_readb_relaxed #endif #ifndef platform_readw_relaxed # define platform_readw_relaxed __ia64_readw_relaxed #endif #ifndef platform_readl_relaxed # define platform_readl_relaxed __ia64_readl_relaxed #endif #ifndef platform_readq_relaxed # define platform_readq_relaxed __ia64_readq_relaxed #endif #ifndef platform_migrate # define platform_migrate machvec_noop_task #endif #ifndef platform_setup_msi_irq # define platform_setup_msi_irq ((ia64_mv_setup_msi_irq_t*)NULL) #endif #ifndef platform_teardown_msi_irq # define platform_teardown_msi_irq ((ia64_mv_teardown_msi_irq_t*)NULL) #endif #ifndef platform_pci_fixup_bus # define platform_pci_fixup_bus machvec_noop_bus #endif #endif /* _ASM_IA64_MACHVEC_H */