Loading drivers/iommu/amd_iommu_init.c +16 −30 Original line number Diff line number Diff line Loading @@ -513,15 +513,12 @@ static int __init find_last_devid_acpi(struct acpi_table_header *table) * write commands to that buffer later and the IOMMU will execute them * asynchronously */ static u8 * __init alloc_command_buffer(struct amd_iommu *iommu) static int __init alloc_command_buffer(struct amd_iommu *iommu) { u8 *cmd_buf = (u8 *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, iommu->cmd_buf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, get_order(CMD_BUFFER_SIZE)); if (cmd_buf == NULL) return NULL; return cmd_buf; return iommu->cmd_buf ? 0 : -ENOMEM; } /* Loading Loading @@ -563,15 +560,12 @@ static void __init free_command_buffer(struct amd_iommu *iommu) } /* allocates the memory where the IOMMU will log its events to */ static u8 * __init alloc_event_buffer(struct amd_iommu *iommu) static int __init alloc_event_buffer(struct amd_iommu *iommu) { iommu->evt_buf = (u8 *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, iommu->evt_buf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, get_order(EVT_BUFFER_SIZE)); if (iommu->evt_buf == NULL) return NULL; return iommu->evt_buf; return iommu->evt_buf ? 0 : -ENOMEM; } static void iommu_enable_event_buffer(struct amd_iommu *iommu) Loading @@ -598,15 +592,12 @@ static void __init free_event_buffer(struct amd_iommu *iommu) } /* allocates the memory where the IOMMU will log its events to */ static u8 * __init alloc_ppr_log(struct amd_iommu *iommu) static int __init alloc_ppr_log(struct amd_iommu *iommu) { iommu->ppr_log = (u8 *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, iommu->ppr_log = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, get_order(PPR_LOG_SIZE)); if (iommu->ppr_log == NULL) return NULL; return iommu->ppr_log; return iommu->ppr_log ? 0 : -ENOMEM; } static void iommu_enable_ppr_log(struct amd_iommu *iommu) Loading Loading @@ -1105,12 +1096,10 @@ static int __init init_iommu_one(struct amd_iommu *iommu, struct ivhd_header *h) if (!iommu->mmio_base) return -ENOMEM; iommu->cmd_buf = alloc_command_buffer(iommu); if (!iommu->cmd_buf) if (alloc_command_buffer(iommu)) return -ENOMEM; iommu->evt_buf = alloc_event_buffer(iommu); if (!iommu->evt_buf) if (alloc_event_buffer(iommu)) return -ENOMEM; iommu->int_enabled = false; Loading Loading @@ -1299,11 +1288,8 @@ static int iommu_init_pci(struct amd_iommu *iommu) amd_iommu_v2_present = true; } if (iommu_feature(iommu, FEATURE_PPR)) { iommu->ppr_log = alloc_ppr_log(iommu); if (!iommu->ppr_log) if (iommu_feature(iommu, FEATURE_PPR) && alloc_ppr_log(iommu)) return -ENOMEM; } if (iommu->cap & (1UL << IOMMU_CAP_NPCACHE)) amd_iommu_np_cache = true; Loading Loading
drivers/iommu/amd_iommu_init.c +16 −30 Original line number Diff line number Diff line Loading @@ -513,15 +513,12 @@ static int __init find_last_devid_acpi(struct acpi_table_header *table) * write commands to that buffer later and the IOMMU will execute them * asynchronously */ static u8 * __init alloc_command_buffer(struct amd_iommu *iommu) static int __init alloc_command_buffer(struct amd_iommu *iommu) { u8 *cmd_buf = (u8 *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, iommu->cmd_buf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, get_order(CMD_BUFFER_SIZE)); if (cmd_buf == NULL) return NULL; return cmd_buf; return iommu->cmd_buf ? 0 : -ENOMEM; } /* Loading Loading @@ -563,15 +560,12 @@ static void __init free_command_buffer(struct amd_iommu *iommu) } /* allocates the memory where the IOMMU will log its events to */ static u8 * __init alloc_event_buffer(struct amd_iommu *iommu) static int __init alloc_event_buffer(struct amd_iommu *iommu) { iommu->evt_buf = (u8 *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, iommu->evt_buf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, get_order(EVT_BUFFER_SIZE)); if (iommu->evt_buf == NULL) return NULL; return iommu->evt_buf; return iommu->evt_buf ? 0 : -ENOMEM; } static void iommu_enable_event_buffer(struct amd_iommu *iommu) Loading @@ -598,15 +592,12 @@ static void __init free_event_buffer(struct amd_iommu *iommu) } /* allocates the memory where the IOMMU will log its events to */ static u8 * __init alloc_ppr_log(struct amd_iommu *iommu) static int __init alloc_ppr_log(struct amd_iommu *iommu) { iommu->ppr_log = (u8 *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, iommu->ppr_log = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, get_order(PPR_LOG_SIZE)); if (iommu->ppr_log == NULL) return NULL; return iommu->ppr_log; return iommu->ppr_log ? 0 : -ENOMEM; } static void iommu_enable_ppr_log(struct amd_iommu *iommu) Loading Loading @@ -1105,12 +1096,10 @@ static int __init init_iommu_one(struct amd_iommu *iommu, struct ivhd_header *h) if (!iommu->mmio_base) return -ENOMEM; iommu->cmd_buf = alloc_command_buffer(iommu); if (!iommu->cmd_buf) if (alloc_command_buffer(iommu)) return -ENOMEM; iommu->evt_buf = alloc_event_buffer(iommu); if (!iommu->evt_buf) if (alloc_event_buffer(iommu)) return -ENOMEM; iommu->int_enabled = false; Loading Loading @@ -1299,11 +1288,8 @@ static int iommu_init_pci(struct amd_iommu *iommu) amd_iommu_v2_present = true; } if (iommu_feature(iommu, FEATURE_PPR)) { iommu->ppr_log = alloc_ppr_log(iommu); if (!iommu->ppr_log) if (iommu_feature(iommu, FEATURE_PPR) && alloc_ppr_log(iommu)) return -ENOMEM; } if (iommu->cap & (1UL << IOMMU_CAP_NPCACHE)) amd_iommu_np_cache = true; Loading