}
 
 struct dma_map_ops dma_iommu_ops = {
-       .alloc_coherent = dma_iommu_alloc_coherent,
-       .free_coherent  = dma_iommu_free_coherent,
-       .map_sg         = dma_iommu_map_sg,
-       .unmap_sg       = dma_iommu_unmap_sg,
-       .dma_supported  = dma_iommu_dma_supported,
-       .map_page       = dma_iommu_map_page,
-       .unmap_page     = dma_iommu_unmap_page,
+       .alloc_coherent         = dma_iommu_alloc_coherent,
+       .free_coherent          = dma_iommu_free_coherent,
+       .map_sg                 = dma_iommu_map_sg,
+       .unmap_sg               = dma_iommu_unmap_sg,
+       .dma_supported          = dma_iommu_dma_supported,
+       .map_page               = dma_iommu_map_page,
+       .unmap_page             = dma_iommu_unmap_page,
        .get_required_mask      = dma_iommu_get_required_mask,
 };
 EXPORT_SYMBOL(dma_iommu_ops);
 
 #endif
 
 struct dma_map_ops dma_direct_ops = {
-       .alloc_coherent = dma_direct_alloc_coherent,
-       .free_coherent  = dma_direct_free_coherent,
-       .map_sg         = dma_direct_map_sg,
-       .unmap_sg       = dma_direct_unmap_sg,
-       .dma_supported  = dma_direct_dma_supported,
-       .map_page       = dma_direct_map_page,
-       .unmap_page     = dma_direct_unmap_page,
-       .get_required_mask      = dma_direct_get_required_mask,
+       .alloc_coherent                 = dma_direct_alloc_coherent,
+       .free_coherent                  = dma_direct_free_coherent,
+       .map_sg                         = dma_direct_map_sg,
+       .unmap_sg                       = dma_direct_unmap_sg,
+       .dma_supported                  = dma_direct_dma_supported,
+       .map_page                       = dma_direct_map_page,
+       .unmap_page                     = dma_direct_unmap_page,
+       .get_required_mask              = dma_direct_get_required_mask,
 #ifdef CONFIG_NOT_COHERENT_CACHE
        .sync_single_for_cpu            = dma_direct_sync_single,
        .sync_single_for_device         = dma_direct_sync_single,
 
 }
 
 static struct dma_map_ops ibmebus_dma_ops = {
-       .alloc_coherent = ibmebus_alloc_coherent,
-       .free_coherent  = ibmebus_free_coherent,
-       .map_sg         = ibmebus_map_sg,
-       .unmap_sg       = ibmebus_unmap_sg,
-       .dma_supported  = ibmebus_dma_supported,
+       .alloc_coherent     = ibmebus_alloc_coherent,
+       .free_coherent      = ibmebus_free_coherent,
+       .map_sg             = ibmebus_map_sg,
+       .unmap_sg           = ibmebus_unmap_sg,
+       .dma_supported      = ibmebus_dma_supported,
        .get_required_mask  = ibmebus_dma_get_required_mask,
-       .map_page       = ibmebus_map_page,
-       .unmap_page     = ibmebus_unmap_page,
+       .map_page           = ibmebus_map_page,
+       .unmap_page         = ibmebus_unmap_page,
 };
 
 static int ibmebus_match_path(struct device *dev, void *data)
 
 }
 
 struct dma_map_ops vio_dma_mapping_ops = {
-       .alloc_coherent = vio_dma_iommu_alloc_coherent,
-       .free_coherent  = vio_dma_iommu_free_coherent,
-       .map_sg         = vio_dma_iommu_map_sg,
-       .unmap_sg       = vio_dma_iommu_unmap_sg,
-       .map_page       = vio_dma_iommu_map_page,
-       .unmap_page     = vio_dma_iommu_unmap_page,
-       .dma_supported  = vio_dma_iommu_dma_supported,
+       .alloc_coherent    = vio_dma_iommu_alloc_coherent,
+       .free_coherent     = vio_dma_iommu_free_coherent,
+       .map_sg            = vio_dma_iommu_map_sg,
+       .unmap_sg          = vio_dma_iommu_unmap_sg,
+       .map_page          = vio_dma_iommu_map_page,
+       .unmap_page        = vio_dma_iommu_unmap_page,
+       .dma_supported     = vio_dma_iommu_dma_supported,
        .get_required_mask = vio_dma_get_required_mask,
 };