| /* |
| * Copyright(c) 2011-2016 Intel Corporation. All rights reserved. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a |
| * copy of this software and associated documentation files (the "Software"), |
| * to deal in the Software without restriction, including without limitation |
| * the rights to use, copy, modify, merge, publish, distribute, sublicense, |
| * and/or sell copies of the Software, and to permit persons to whom the |
| * Software is furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice (including the next |
| * paragraph) shall be included in all copies or substantial portions of the |
| * Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
| * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
| * SOFTWARE. |
| * |
| * Authors: |
| * Kevin Tian <kevin.tian@intel.com> |
| * Eddie Dong <eddie.dong@intel.com> |
| * |
| * Contributors: |
| * Niu Bing <bing.niu@intel.com> |
| * Zhi Wang <zhi.a.wang@intel.com> |
| * |
| */ |
| |
| #include <linux/types.h> |
| #include <xen/xen.h> |
| #include <linux/kthread.h> |
| |
| #include "i915_drv.h" |
| #include "gvt.h" |
| #include <linux/vfio.h> |
| #include <linux/mdev.h> |
| |
| struct intel_gvt_host intel_gvt_host; |
| |
| static const char * const supported_hypervisors[] = { |
| [INTEL_GVT_HYPERVISOR_XEN] = "XEN", |
| [INTEL_GVT_HYPERVISOR_KVM] = "KVM", |
| }; |
| |
| static struct intel_vgpu_type *intel_gvt_find_vgpu_type(struct intel_gvt *gvt, |
| const char *name) |
| { |
| int i; |
| struct intel_vgpu_type *t; |
| const char *driver_name = dev_driver_string( |
| &gvt->dev_priv->drm.pdev->dev); |
| |
| for (i = 0; i < gvt->num_types; i++) { |
| t = &gvt->types[i]; |
| if (!strncmp(t->name, name + strlen(driver_name) + 1, |
| sizeof(t->name))) |
| return t; |
| } |
| |
| return NULL; |
| } |
| |
| static ssize_t available_instances_show(struct kobject *kobj, |
| struct device *dev, char *buf) |
| { |
| struct intel_vgpu_type *type; |
| unsigned int num = 0; |
| void *gvt = kdev_to_i915(dev)->gvt; |
| |
| type = intel_gvt_find_vgpu_type(gvt, kobject_name(kobj)); |
| if (!type) |
| num = 0; |
| else |
| num = type->avail_instance; |
| |
| return sprintf(buf, "%u\n", num); |
| } |
| |
| static ssize_t device_api_show(struct kobject *kobj, struct device *dev, |
| char *buf) |
| { |
| return sprintf(buf, "%s\n", VFIO_DEVICE_API_PCI_STRING); |
| } |
| |
| static ssize_t description_show(struct kobject *kobj, struct device *dev, |
| char *buf) |
| { |
| struct intel_vgpu_type *type; |
| void *gvt = kdev_to_i915(dev)->gvt; |
| |
| type = intel_gvt_find_vgpu_type(gvt, kobject_name(kobj)); |
| if (!type) |
| return 0; |
| |
| return sprintf(buf, "low_gm_size: %dMB\nhigh_gm_size: %dMB\n" |
| "fence: %d\nresolution: %s\n" |
| "weight: %d\n", |
| BYTES_TO_MB(type->low_gm_size), |
| BYTES_TO_MB(type->high_gm_size), |
| type->fence, vgpu_edid_str(type->resolution), |
| type->weight); |
| } |
| |
| static MDEV_TYPE_ATTR_RO(available_instances); |
| static MDEV_TYPE_ATTR_RO(device_api); |
| static MDEV_TYPE_ATTR_RO(description); |
| |
| static struct attribute *gvt_type_attrs[] = { |
| &mdev_type_attr_available_instances.attr, |
| &mdev_type_attr_device_api.attr, |
| &mdev_type_attr_description.attr, |
| NULL, |
| }; |
| |
| static struct attribute_group *gvt_vgpu_type_groups[] = { |
| [0 ... NR_MAX_INTEL_VGPU_TYPES - 1] = NULL, |
| }; |
| |
| static bool intel_get_gvt_attrs(struct attribute ***type_attrs, |
| struct attribute_group ***intel_vgpu_type_groups) |
| { |
| *type_attrs = gvt_type_attrs; |
| *intel_vgpu_type_groups = gvt_vgpu_type_groups; |
| return true; |
| } |
| |
| static bool intel_gvt_init_vgpu_type_groups(struct intel_gvt *gvt) |
| { |
| int i, j; |
| struct intel_vgpu_type *type; |
| struct attribute_group *group; |
| |
| for (i = 0; i < gvt->num_types; i++) { |
| type = &gvt->types[i]; |
| |
| group = kzalloc(sizeof(struct attribute_group), GFP_KERNEL); |
| if (WARN_ON(!group)) |
| goto unwind; |
| |
| group->name = type->name; |
| group->attrs = gvt_type_attrs; |
| gvt_vgpu_type_groups[i] = group; |
| } |
| |
| return true; |
| |
| unwind: |
| for (j = 0; j < i; j++) { |
| group = gvt_vgpu_type_groups[j]; |
| kfree(group); |
| } |
| |
| return false; |
| } |
| |
| static void intel_gvt_cleanup_vgpu_type_groups(struct intel_gvt *gvt) |
| { |
| int i; |
| struct attribute_group *group; |
| |
| for (i = 0; i < gvt->num_types; i++) { |
| group = gvt_vgpu_type_groups[i]; |
| gvt_vgpu_type_groups[i] = NULL; |
| kfree(group); |
| } |
| } |
| |
| static const struct intel_gvt_ops intel_gvt_ops = { |
| .emulate_cfg_read = intel_vgpu_emulate_cfg_read, |
| .emulate_cfg_write = intel_vgpu_emulate_cfg_write, |
| .emulate_mmio_read = intel_vgpu_emulate_mmio_read, |
| .emulate_mmio_write = intel_vgpu_emulate_mmio_write, |
| .vgpu_create = intel_gvt_create_vgpu, |
| .vgpu_destroy = intel_gvt_destroy_vgpu, |
| .vgpu_reset = intel_gvt_reset_vgpu, |
| .vgpu_activate = intel_gvt_activate_vgpu, |
| .vgpu_deactivate = intel_gvt_deactivate_vgpu, |
| .gvt_find_vgpu_type = intel_gvt_find_vgpu_type, |
| .get_gvt_attrs = intel_get_gvt_attrs, |
| .vgpu_query_plane = intel_vgpu_query_plane, |
| .vgpu_get_dmabuf = intel_vgpu_get_dmabuf, |
| .write_protect_handler = intel_vgpu_page_track_handler, |
| }; |
| |
| /** |
| * intel_gvt_init_host - Load MPT modules and detect if we're running in host |
| * @gvt: intel gvt device |
| * |
| * This function is called at the driver loading stage. If failed to find a |
| * loadable MPT module or detect currently we're running in a VM, then GVT-g |
| * will be disabled |
| * |
| * Returns: |
| * Zero on success, negative error code if failed. |
| * |
| */ |
| int intel_gvt_init_host(void) |
| { |
| if (intel_gvt_host.initialized) |
| return 0; |
| |
| /* Xen DOM U */ |
| if (xen_domain() && !xen_initial_domain()) |
| return -ENODEV; |
| |
| /* Try to load MPT modules for hypervisors */ |
| if (xen_initial_domain()) { |
| /* In Xen dom0 */ |
| intel_gvt_host.mpt = try_then_request_module( |
| symbol_get(xengt_mpt), "xengt"); |
| intel_gvt_host.hypervisor_type = INTEL_GVT_HYPERVISOR_XEN; |
| } else { |
| #if IS_ENABLED(CONFIG_DRM_I915_GVT_KVMGT) |
| /* not in Xen. Try KVMGT */ |
| intel_gvt_host.mpt = try_then_request_module( |
| symbol_get(kvmgt_mpt), "kvmgt"); |
| intel_gvt_host.hypervisor_type = INTEL_GVT_HYPERVISOR_KVM; |
| #endif |
| } |
| |
| /* Fail to load MPT modules - bail out */ |
| if (!intel_gvt_host.mpt) |
| return -EINVAL; |
| |
| gvt_dbg_core("Running with hypervisor %s in host mode\n", |
| supported_hypervisors[intel_gvt_host.hypervisor_type]); |
| |
| intel_gvt_host.initialized = true; |
| return 0; |
| } |
| |
| static void init_device_info(struct intel_gvt *gvt) |
| { |
| struct intel_gvt_device_info *info = &gvt->device_info; |
| struct pci_dev *pdev = gvt->dev_priv->drm.pdev; |
| |
| if (IS_BROADWELL(gvt->dev_priv) || IS_SKYLAKE(gvt->dev_priv) |
| || IS_KABYLAKE(gvt->dev_priv)) { |
| info->max_support_vgpus = 8; |
| info->cfg_space_size = PCI_CFG_SPACE_EXP_SIZE; |
| info->mmio_size = 2 * 1024 * 1024; |
| info->mmio_bar = 0; |
| info->gtt_start_offset = 8 * 1024 * 1024; |
| info->gtt_entry_size = 8; |
| info->gtt_entry_size_shift = 3; |
| info->gmadr_bytes_in_cmd = 8; |
| info->max_surface_size = 36 * 1024 * 1024; |
| } |
| info->msi_cap_offset = pdev->msi_cap; |
| } |
| |
| static int gvt_service_thread(void *data) |
| { |
| struct intel_gvt *gvt = (struct intel_gvt *)data; |
| int ret; |
| |
| gvt_dbg_core("service thread start\n"); |
| |
| while (!kthread_should_stop()) { |
| ret = wait_event_interruptible(gvt->service_thread_wq, |
| kthread_should_stop() || gvt->service_request); |
| |
| if (kthread_should_stop()) |
| break; |
| |
| if (WARN_ONCE(ret, "service thread is waken up by signal.\n")) |
| continue; |
| |
| if (test_and_clear_bit(INTEL_GVT_REQUEST_EMULATE_VBLANK, |
| (void *)&gvt->service_request)) { |
| mutex_lock(&gvt->lock); |
| intel_gvt_emulate_vblank(gvt); |
| mutex_unlock(&gvt->lock); |
| } |
| |
| if (test_bit(INTEL_GVT_REQUEST_SCHED, |
| (void *)&gvt->service_request) || |
| test_bit(INTEL_GVT_REQUEST_EVENT_SCHED, |
| (void *)&gvt->service_request)) { |
| intel_gvt_schedule(gvt); |
| } |
| } |
| |
| return 0; |
| } |
| |
| static void clean_service_thread(struct intel_gvt *gvt) |
| { |
| kthread_stop(gvt->service_thread); |
| } |
| |
| static int init_service_thread(struct intel_gvt *gvt) |
| { |
| init_waitqueue_head(&gvt->service_thread_wq); |
| |
| gvt->service_thread = kthread_run(gvt_service_thread, |
| gvt, "gvt_service_thread"); |
| if (IS_ERR(gvt->service_thread)) { |
| gvt_err("fail to start service thread.\n"); |
| return PTR_ERR(gvt->service_thread); |
| } |
| return 0; |
| } |
| |
| /** |
| * intel_gvt_clean_device - clean a GVT device |
| * @gvt: intel gvt device |
| * |
| * This function is called at the driver unloading stage, to free the |
| * resources owned by a GVT device. |
| * |
| */ |
| void intel_gvt_clean_device(struct drm_i915_private *dev_priv) |
| { |
| struct intel_gvt *gvt = to_gvt(dev_priv); |
| |
| if (WARN_ON(!gvt)) |
| return; |
| |
| intel_gvt_debugfs_clean(gvt); |
| clean_service_thread(gvt); |
| intel_gvt_clean_cmd_parser(gvt); |
| intel_gvt_clean_sched_policy(gvt); |
| intel_gvt_clean_workload_scheduler(gvt); |
| intel_gvt_clean_gtt(gvt); |
| intel_gvt_clean_irq(gvt); |
| intel_gvt_clean_mmio_info(gvt); |
| intel_gvt_free_firmware(gvt); |
| |
| intel_gvt_hypervisor_host_exit(&dev_priv->drm.pdev->dev, gvt); |
| intel_gvt_cleanup_vgpu_type_groups(gvt); |
| intel_gvt_clean_vgpu_types(gvt); |
| |
| idr_destroy(&gvt->vgpu_idr); |
| |
| intel_gvt_destroy_idle_vgpu(gvt->idle_vgpu); |
| |
| kfree(dev_priv->gvt); |
| dev_priv->gvt = NULL; |
| } |
| |
| /** |
| * intel_gvt_init_device - initialize a GVT device |
| * @dev_priv: drm i915 private data |
| * |
| * This function is called at the initialization stage, to initialize |
| * necessary GVT components. |
| * |
| * Returns: |
| * Zero on success, negative error code if failed. |
| * |
| */ |
| int intel_gvt_init_device(struct drm_i915_private *dev_priv) |
| { |
| struct intel_gvt *gvt; |
| struct intel_vgpu *vgpu; |
| int ret; |
| |
| /* |
| * Cannot initialize GVT device without intel_gvt_host gets |
| * initialized first. |
| */ |
| if (WARN_ON(!intel_gvt_host.initialized)) |
| return -EINVAL; |
| |
| if (WARN_ON(dev_priv->gvt)) |
| return -EEXIST; |
| |
| gvt = kzalloc(sizeof(struct intel_gvt), GFP_KERNEL); |
| if (!gvt) |
| return -ENOMEM; |
| |
| gvt_dbg_core("init gvt device\n"); |
| |
| idr_init(&gvt->vgpu_idr); |
| spin_lock_init(&gvt->scheduler.mmio_context_lock); |
| mutex_init(&gvt->lock); |
| gvt->dev_priv = dev_priv; |
| |
| init_device_info(gvt); |
| |
| ret = intel_gvt_setup_mmio_info(gvt); |
| if (ret) |
| goto out_clean_idr; |
| |
| intel_gvt_init_engine_mmio_context(gvt); |
| |
| ret = intel_gvt_load_firmware(gvt); |
| if (ret) |
| goto out_clean_mmio_info; |
| |
| ret = intel_gvt_init_irq(gvt); |
| if (ret) |
| goto out_free_firmware; |
| |
| ret = intel_gvt_init_gtt(gvt); |
| if (ret) |
| goto out_clean_irq; |
| |
| ret = intel_gvt_init_workload_scheduler(gvt); |
| if (ret) |
| goto out_clean_gtt; |
| |
| ret = intel_gvt_init_sched_policy(gvt); |
| if (ret) |
| goto out_clean_workload_scheduler; |
| |
| ret = intel_gvt_init_cmd_parser(gvt); |
| if (ret) |
| goto out_clean_sched_policy; |
| |
| ret = init_service_thread(gvt); |
| if (ret) |
| goto out_clean_cmd_parser; |
| |
| ret = intel_gvt_init_vgpu_types(gvt); |
| if (ret) |
| goto out_clean_thread; |
| |
| ret = intel_gvt_init_vgpu_type_groups(gvt); |
| if (ret == false) { |
| gvt_err("failed to init vgpu type groups: %d\n", ret); |
| goto out_clean_types; |
| } |
| |
| ret = intel_gvt_hypervisor_host_init(&dev_priv->drm.pdev->dev, gvt, |
| &intel_gvt_ops); |
| if (ret) { |
| gvt_err("failed to register gvt-g host device: %d\n", ret); |
| goto out_clean_types; |
| } |
| |
| vgpu = intel_gvt_create_idle_vgpu(gvt); |
| if (IS_ERR(vgpu)) { |
| ret = PTR_ERR(vgpu); |
| gvt_err("failed to create idle vgpu\n"); |
| goto out_clean_types; |
| } |
| gvt->idle_vgpu = vgpu; |
| |
| ret = intel_gvt_debugfs_init(gvt); |
| if (ret) |
| gvt_err("debugfs registeration failed, go on.\n"); |
| |
| gvt_dbg_core("gvt device initialization is done\n"); |
| dev_priv->gvt = gvt; |
| return 0; |
| |
| out_clean_types: |
| intel_gvt_clean_vgpu_types(gvt); |
| out_clean_thread: |
| clean_service_thread(gvt); |
| out_clean_cmd_parser: |
| intel_gvt_clean_cmd_parser(gvt); |
| out_clean_sched_policy: |
| intel_gvt_clean_sched_policy(gvt); |
| out_clean_workload_scheduler: |
| intel_gvt_clean_workload_scheduler(gvt); |
| out_clean_gtt: |
| intel_gvt_clean_gtt(gvt); |
| out_clean_irq: |
| intel_gvt_clean_irq(gvt); |
| out_free_firmware: |
| intel_gvt_free_firmware(gvt); |
| out_clean_mmio_info: |
| intel_gvt_clean_mmio_info(gvt); |
| out_clean_idr: |
| idr_destroy(&gvt->vgpu_idr); |
| kfree(gvt); |
| return ret; |
| } |