linux/drivers/gpu/drm/i915/gvt/gvt.c
<<
>>
Prefs
   1/*
   2 * Copyright(c) 2011-2016 Intel Corporation. All rights reserved.
   3 *
   4 * Permission is hereby granted, free of charge, to any person obtaining a
   5 * copy of this software and associated documentation files (the "Software"),
   6 * to deal in the Software without restriction, including without limitation
   7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
   8 * and/or sell copies of the Software, and to permit persons to whom the
   9 * Software is furnished to do so, subject to the following conditions:
  10 *
  11 * The above copyright notice and this permission notice (including the next
  12 * paragraph) shall be included in all copies or substantial portions of the
  13 * Software.
  14 *
  15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
  18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
  20 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
  21 * SOFTWARE.
  22 *
  23 * Authors:
  24 *    Kevin Tian <kevin.tian@intel.com>
  25 *    Eddie Dong <eddie.dong@intel.com>
  26 *
  27 * Contributors:
  28 *    Niu Bing <bing.niu@intel.com>
  29 *    Zhi Wang <zhi.a.wang@intel.com>
  30 *
  31 */
  32
  33#include <linux/types.h>
  34#include <linux/kthread.h>
  35
  36#include "i915_drv.h"
  37#include "intel_gvt.h"
  38#include "gvt.h"
  39#include <linux/vfio.h>
  40#include <linux/mdev.h>
  41
  42struct intel_gvt_host intel_gvt_host;
  43
  44static const char * const supported_hypervisors[] = {
  45        [INTEL_GVT_HYPERVISOR_XEN] = "XEN",
  46        [INTEL_GVT_HYPERVISOR_KVM] = "KVM",
  47};
  48
  49static struct intel_vgpu_type *intel_gvt_find_vgpu_type(struct intel_gvt *gvt,
  50                const char *name)
  51{
  52        const char *driver_name =
  53                dev_driver_string(&gvt->gt->i915->drm.pdev->dev);
  54        int i;
  55
  56        name += strlen(driver_name) + 1;
  57        for (i = 0; i < gvt->num_types; i++) {
  58                struct intel_vgpu_type *t = &gvt->types[i];
  59
  60                if (!strncmp(t->name, name, sizeof(t->name)))
  61                        return t;
  62        }
  63
  64        return NULL;
  65}
  66
  67static ssize_t available_instances_show(struct kobject *kobj,
  68                                        struct device *dev, char *buf)
  69{
  70        struct intel_vgpu_type *type;
  71        unsigned int num = 0;
  72        void *gvt = kdev_to_i915(dev)->gvt;
  73
  74        type = intel_gvt_find_vgpu_type(gvt, kobject_name(kobj));
  75        if (!type)
  76                num = 0;
  77        else
  78                num = type->avail_instance;
  79
  80        return sprintf(buf, "%u\n", num);
  81}
  82
  83static ssize_t device_api_show(struct kobject *kobj, struct device *dev,
  84                char *buf)
  85{
  86        return sprintf(buf, "%s\n", VFIO_DEVICE_API_PCI_STRING);
  87}
  88
  89static ssize_t description_show(struct kobject *kobj, struct device *dev,
  90                char *buf)
  91{
  92        struct intel_vgpu_type *type;
  93        void *gvt = kdev_to_i915(dev)->gvt;
  94
  95        type = intel_gvt_find_vgpu_type(gvt, kobject_name(kobj));
  96        if (!type)
  97                return 0;
  98
  99        return sprintf(buf, "low_gm_size: %dMB\nhigh_gm_size: %dMB\n"
 100                       "fence: %d\nresolution: %s\n"
 101                       "weight: %d\n",
 102                       BYTES_TO_MB(type->low_gm_size),
 103                       BYTES_TO_MB(type->high_gm_size),
 104                       type->fence, vgpu_edid_str(type->resolution),
 105                       type->weight);
 106}
 107
 108static MDEV_TYPE_ATTR_RO(available_instances);
 109static MDEV_TYPE_ATTR_RO(device_api);
 110static MDEV_TYPE_ATTR_RO(description);
 111
 112static struct attribute *gvt_type_attrs[] = {
 113        &mdev_type_attr_available_instances.attr,
 114        &mdev_type_attr_device_api.attr,
 115        &mdev_type_attr_description.attr,
 116        NULL,
 117};
 118
 119static struct attribute_group *gvt_vgpu_type_groups[] = {
 120        [0 ... NR_MAX_INTEL_VGPU_TYPES - 1] = NULL,
 121};
 122
 123static bool intel_get_gvt_attrs(struct attribute_group ***intel_vgpu_type_groups)
 124{
 125        *intel_vgpu_type_groups = gvt_vgpu_type_groups;
 126        return true;
 127}
 128
 129static int intel_gvt_init_vgpu_type_groups(struct intel_gvt *gvt)
 130{
 131        int i, j;
 132        struct intel_vgpu_type *type;
 133        struct attribute_group *group;
 134
 135        for (i = 0; i < gvt->num_types; i++) {
 136                type = &gvt->types[i];
 137
 138                group = kzalloc(sizeof(struct attribute_group), GFP_KERNEL);
 139                if (WARN_ON(!group))
 140                        goto unwind;
 141
 142                group->name = type->name;
 143                group->attrs = gvt_type_attrs;
 144                gvt_vgpu_type_groups[i] = group;
 145        }
 146
 147        return 0;
 148
 149unwind:
 150        for (j = 0; j < i; j++) {
 151                group = gvt_vgpu_type_groups[j];
 152                kfree(group);
 153        }
 154
 155        return -ENOMEM;
 156}
 157
 158static void intel_gvt_cleanup_vgpu_type_groups(struct intel_gvt *gvt)
 159{
 160        int i;
 161        struct attribute_group *group;
 162
 163        for (i = 0; i < gvt->num_types; i++) {
 164                group = gvt_vgpu_type_groups[i];
 165                gvt_vgpu_type_groups[i] = NULL;
 166                kfree(group);
 167        }
 168}
 169
 170static const struct intel_gvt_ops intel_gvt_ops = {
 171        .emulate_cfg_read = intel_vgpu_emulate_cfg_read,
 172        .emulate_cfg_write = intel_vgpu_emulate_cfg_write,
 173        .emulate_mmio_read = intel_vgpu_emulate_mmio_read,
 174        .emulate_mmio_write = intel_vgpu_emulate_mmio_write,
 175        .vgpu_create = intel_gvt_create_vgpu,
 176        .vgpu_destroy = intel_gvt_destroy_vgpu,
 177        .vgpu_release = intel_gvt_release_vgpu,
 178        .vgpu_reset = intel_gvt_reset_vgpu,
 179        .vgpu_activate = intel_gvt_activate_vgpu,
 180        .vgpu_deactivate = intel_gvt_deactivate_vgpu,
 181        .gvt_find_vgpu_type = intel_gvt_find_vgpu_type,
 182        .get_gvt_attrs = intel_get_gvt_attrs,
 183        .vgpu_query_plane = intel_vgpu_query_plane,
 184        .vgpu_get_dmabuf = intel_vgpu_get_dmabuf,
 185        .write_protect_handler = intel_vgpu_page_track_handler,
 186        .emulate_hotplug = intel_vgpu_emulate_hotplug,
 187};
 188
 189static void init_device_info(struct intel_gvt *gvt)
 190{
 191        struct intel_gvt_device_info *info = &gvt->device_info;
 192        struct pci_dev *pdev = gvt->gt->i915->drm.pdev;
 193
 194        info->max_support_vgpus = 8;
 195        info->cfg_space_size = PCI_CFG_SPACE_EXP_SIZE;
 196        info->mmio_size = 2 * 1024 * 1024;
 197        info->mmio_bar = 0;
 198        info->gtt_start_offset = 8 * 1024 * 1024;
 199        info->gtt_entry_size = 8;
 200        info->gtt_entry_size_shift = 3;
 201        info->gmadr_bytes_in_cmd = 8;
 202        info->max_surface_size = 36 * 1024 * 1024;
 203        info->msi_cap_offset = pdev->msi_cap;
 204}
 205
 206static int gvt_service_thread(void *data)
 207{
 208        struct intel_gvt *gvt = (struct intel_gvt *)data;
 209        int ret;
 210
 211        gvt_dbg_core("service thread start\n");
 212
 213        while (!kthread_should_stop()) {
 214                ret = wait_event_interruptible(gvt->service_thread_wq,
 215                                kthread_should_stop() || gvt->service_request);
 216
 217                if (kthread_should_stop())
 218                        break;
 219
 220                if (WARN_ONCE(ret, "service thread is waken up by signal.\n"))
 221                        continue;
 222
 223                if (test_and_clear_bit(INTEL_GVT_REQUEST_EMULATE_VBLANK,
 224                                        (void *)&gvt->service_request))
 225                        intel_gvt_emulate_vblank(gvt);
 226
 227                if (test_bit(INTEL_GVT_REQUEST_SCHED,
 228                                (void *)&gvt->service_request) ||
 229                        test_bit(INTEL_GVT_REQUEST_EVENT_SCHED,
 230                                        (void *)&gvt->service_request)) {
 231                        intel_gvt_schedule(gvt);
 232                }
 233        }
 234
 235        return 0;
 236}
 237
 238static void clean_service_thread(struct intel_gvt *gvt)
 239{
 240        kthread_stop(gvt->service_thread);
 241}
 242
 243static int init_service_thread(struct intel_gvt *gvt)
 244{
 245        init_waitqueue_head(&gvt->service_thread_wq);
 246
 247        gvt->service_thread = kthread_run(gvt_service_thread,
 248                        gvt, "gvt_service_thread");
 249        if (IS_ERR(gvt->service_thread)) {
 250                gvt_err("fail to start service thread.\n");
 251                return PTR_ERR(gvt->service_thread);
 252        }
 253        return 0;
 254}
 255
 256/**
 257 * intel_gvt_clean_device - clean a GVT device
 258 * @i915: i915 private
 259 *
 260 * This function is called at the driver unloading stage, to free the
 261 * resources owned by a GVT device.
 262 *
 263 */
 264void intel_gvt_clean_device(struct drm_i915_private *i915)
 265{
 266        struct intel_gvt *gvt = fetch_and_zero(&i915->gvt);
 267
 268        if (drm_WARN_ON(&i915->drm, !gvt))
 269                return;
 270
 271        intel_gvt_destroy_idle_vgpu(gvt->idle_vgpu);
 272        intel_gvt_cleanup_vgpu_type_groups(gvt);
 273        intel_gvt_clean_vgpu_types(gvt);
 274
 275        intel_gvt_debugfs_clean(gvt);
 276        clean_service_thread(gvt);
 277        intel_gvt_clean_cmd_parser(gvt);
 278        intel_gvt_clean_sched_policy(gvt);
 279        intel_gvt_clean_workload_scheduler(gvt);
 280        intel_gvt_clean_gtt(gvt);
 281        intel_gvt_clean_irq(gvt);
 282        intel_gvt_free_firmware(gvt);
 283        intel_gvt_clean_mmio_info(gvt);
 284        idr_destroy(&gvt->vgpu_idr);
 285
 286        kfree(i915->gvt);
 287}
 288
 289/**
 290 * intel_gvt_init_device - initialize a GVT device
 291 * @i915: drm i915 private data
 292 *
 293 * This function is called at the initialization stage, to initialize
 294 * necessary GVT components.
 295 *
 296 * Returns:
 297 * Zero on success, negative error code if failed.
 298 *
 299 */
 300int intel_gvt_init_device(struct drm_i915_private *i915)
 301{
 302        struct intel_gvt *gvt;
 303        struct intel_vgpu *vgpu;
 304        int ret;
 305
 306        if (drm_WARN_ON(&i915->drm, i915->gvt))
 307                return -EEXIST;
 308
 309        gvt = kzalloc(sizeof(struct intel_gvt), GFP_KERNEL);
 310        if (!gvt)
 311                return -ENOMEM;
 312
 313        gvt_dbg_core("init gvt device\n");
 314
 315        idr_init_base(&gvt->vgpu_idr, 1);
 316        spin_lock_init(&gvt->scheduler.mmio_context_lock);
 317        mutex_init(&gvt->lock);
 318        mutex_init(&gvt->sched_lock);
 319        gvt->gt = &i915->gt;
 320        i915->gvt = gvt;
 321
 322        init_device_info(gvt);
 323
 324        ret = intel_gvt_setup_mmio_info(gvt);
 325        if (ret)
 326                goto out_clean_idr;
 327
 328        intel_gvt_init_engine_mmio_context(gvt);
 329
 330        ret = intel_gvt_load_firmware(gvt);
 331        if (ret)
 332                goto out_clean_mmio_info;
 333
 334        ret = intel_gvt_init_irq(gvt);
 335        if (ret)
 336                goto out_free_firmware;
 337
 338        ret = intel_gvt_init_gtt(gvt);
 339        if (ret)
 340                goto out_clean_irq;
 341
 342        ret = intel_gvt_init_workload_scheduler(gvt);
 343        if (ret)
 344                goto out_clean_gtt;
 345
 346        ret = intel_gvt_init_sched_policy(gvt);
 347        if (ret)
 348                goto out_clean_workload_scheduler;
 349
 350        ret = intel_gvt_init_cmd_parser(gvt);
 351        if (ret)
 352                goto out_clean_sched_policy;
 353
 354        ret = init_service_thread(gvt);
 355        if (ret)
 356                goto out_clean_cmd_parser;
 357
 358        ret = intel_gvt_init_vgpu_types(gvt);
 359        if (ret)
 360                goto out_clean_thread;
 361
 362        ret = intel_gvt_init_vgpu_type_groups(gvt);
 363        if (ret) {
 364                gvt_err("failed to init vgpu type groups: %d\n", ret);
 365                goto out_clean_types;
 366        }
 367
 368        vgpu = intel_gvt_create_idle_vgpu(gvt);
 369        if (IS_ERR(vgpu)) {
 370                ret = PTR_ERR(vgpu);
 371                gvt_err("failed to create idle vgpu\n");
 372                goto out_clean_types;
 373        }
 374        gvt->idle_vgpu = vgpu;
 375
 376        intel_gvt_debugfs_init(gvt);
 377
 378        gvt_dbg_core("gvt device initialization is done\n");
 379        intel_gvt_host.dev = &i915->drm.pdev->dev;
 380        intel_gvt_host.initialized = true;
 381        return 0;
 382
 383out_clean_types:
 384        intel_gvt_clean_vgpu_types(gvt);
 385out_clean_thread:
 386        clean_service_thread(gvt);
 387out_clean_cmd_parser:
 388        intel_gvt_clean_cmd_parser(gvt);
 389out_clean_sched_policy:
 390        intel_gvt_clean_sched_policy(gvt);
 391out_clean_workload_scheduler:
 392        intel_gvt_clean_workload_scheduler(gvt);
 393out_clean_gtt:
 394        intel_gvt_clean_gtt(gvt);
 395out_clean_irq:
 396        intel_gvt_clean_irq(gvt);
 397out_free_firmware:
 398        intel_gvt_free_firmware(gvt);
 399out_clean_mmio_info:
 400        intel_gvt_clean_mmio_info(gvt);
 401out_clean_idr:
 402        idr_destroy(&gvt->vgpu_idr);
 403        kfree(gvt);
 404        i915->gvt = NULL;
 405        return ret;
 406}
 407
 408int
 409intel_gvt_pm_resume(struct intel_gvt *gvt)
 410{
 411        intel_gvt_restore_fence(gvt);
 412        intel_gvt_restore_mmio(gvt);
 413        intel_gvt_restore_ggtt(gvt);
 414        return 0;
 415}
 416
 417int
 418intel_gvt_register_hypervisor(const struct intel_gvt_mpt *m)
 419{
 420        int ret;
 421        void *gvt;
 422
 423        if (!intel_gvt_host.initialized)
 424                return -ENODEV;
 425
 426        if (m->type != INTEL_GVT_HYPERVISOR_KVM &&
 427            m->type != INTEL_GVT_HYPERVISOR_XEN)
 428                return -EINVAL;
 429
 430        /* Get a reference for device model module */
 431        if (!try_module_get(THIS_MODULE))
 432                return -ENODEV;
 433
 434        intel_gvt_host.mpt = m;
 435        intel_gvt_host.hypervisor_type = m->type;
 436        gvt = (void *)kdev_to_i915(intel_gvt_host.dev)->gvt;
 437
 438        ret = intel_gvt_hypervisor_host_init(intel_gvt_host.dev, gvt,
 439                                             &intel_gvt_ops);
 440        if (ret < 0) {
 441                gvt_err("Failed to init %s hypervisor module\n",
 442                        supported_hypervisors[intel_gvt_host.hypervisor_type]);
 443                module_put(THIS_MODULE);
 444                return -ENODEV;
 445        }
 446        gvt_dbg_core("Running with hypervisor %s in host mode\n",
 447                     supported_hypervisors[intel_gvt_host.hypervisor_type]);
 448        return 0;
 449}
 450EXPORT_SYMBOL_GPL(intel_gvt_register_hypervisor);
 451
 452void
 453intel_gvt_unregister_hypervisor(void)
 454{
 455        intel_gvt_hypervisor_host_exit(intel_gvt_host.dev);
 456        module_put(THIS_MODULE);
 457}
 458EXPORT_SYMBOL_GPL(intel_gvt_unregister_hypervisor);
 459