linux/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
<<
>>
Prefs
   1/*
   2 * Copyright 2016 Advanced Micro Devices, Inc.
   3 *
   4 * Permission is hereby granted, free of charge, to any person obtaining a
   5 * copy of this software and associated documentation files (the "Software"),
   6 * to deal in the Software without restriction, including without limitation
   7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
   8 * and/or sell copies of the Software, and to permit persons to whom the
   9 * Software is furnished to do so, subject to the following conditions:
  10 *
  11 * The above copyright notice and this permission notice shall be included in
  12 * all copies or substantial portions of the Software.
  13 *
  14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
  17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
  18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
  19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
  20 * OTHER DEALINGS IN THE SOFTWARE.
  21 *
  22 * Authors: Christian König
  23 */
  24
  25#include <drm/drmP.h>
  26#include "amdgpu.h"
  27
  28struct amdgpu_vram_mgr {
  29        struct drm_mm mm;
  30        spinlock_t lock;
  31        atomic64_t usage;
  32        atomic64_t vis_usage;
  33};
  34
  35/**
  36 * amdgpu_vram_mgr_init - init VRAM manager and DRM MM
  37 *
  38 * @man: TTM memory type manager
  39 * @p_size: maximum size of VRAM
  40 *
  41 * Allocate and initialize the VRAM manager.
  42 */
  43static int amdgpu_vram_mgr_init(struct ttm_mem_type_manager *man,
  44                                unsigned long p_size)
  45{
  46        struct amdgpu_vram_mgr *mgr;
  47
  48        mgr = kzalloc(sizeof(*mgr), GFP_KERNEL);
  49        if (!mgr)
  50                return -ENOMEM;
  51
  52        drm_mm_init(&mgr->mm, 0, p_size);
  53        spin_lock_init(&mgr->lock);
  54        man->priv = mgr;
  55        return 0;
  56}
  57
  58/**
  59 * amdgpu_vram_mgr_fini - free and destroy VRAM manager
  60 *
  61 * @man: TTM memory type manager
  62 *
  63 * Destroy and free the VRAM manager, returns -EBUSY if ranges are still
  64 * allocated inside it.
  65 */
  66static int amdgpu_vram_mgr_fini(struct ttm_mem_type_manager *man)
  67{
  68        struct amdgpu_vram_mgr *mgr = man->priv;
  69
  70        spin_lock(&mgr->lock);
  71        drm_mm_takedown(&mgr->mm);
  72        spin_unlock(&mgr->lock);
  73        kfree(mgr);
  74        man->priv = NULL;
  75        return 0;
  76}
  77
  78/**
  79 * amdgpu_vram_mgr_vis_size - Calculate visible node size
  80 *
  81 * @adev: amdgpu device structure
  82 * @node: MM node structure
  83 *
  84 * Calculate how many bytes of the MM node are inside visible VRAM
  85 */
  86static u64 amdgpu_vram_mgr_vis_size(struct amdgpu_device *adev,
  87                                    struct drm_mm_node *node)
  88{
  89        uint64_t start = node->start << PAGE_SHIFT;
  90        uint64_t end = (node->size + node->start) << PAGE_SHIFT;
  91
  92        if (start >= adev->gmc.visible_vram_size)
  93                return 0;
  94
  95        return (end > adev->gmc.visible_vram_size ?
  96                adev->gmc.visible_vram_size : end) - start;
  97}
  98
  99/**
 100 * amdgpu_vram_mgr_bo_visible_size - CPU visible BO size
 101 *
 102 * @bo: &amdgpu_bo buffer object (must be in VRAM)
 103 *
 104 * Returns:
 105 * How much of the given &amdgpu_bo buffer object lies in CPU visible VRAM.
 106 */
 107u64 amdgpu_vram_mgr_bo_visible_size(struct amdgpu_bo *bo)
 108{
 109        struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
 110        struct ttm_mem_reg *mem = &bo->tbo.mem;
 111        struct drm_mm_node *nodes = mem->mm_node;
 112        unsigned pages = mem->num_pages;
 113        u64 usage;
 114
 115        if (amdgpu_gmc_vram_full_visible(&adev->gmc))
 116                return amdgpu_bo_size(bo);
 117
 118        if (mem->start >= adev->gmc.visible_vram_size >> PAGE_SHIFT)
 119                return 0;
 120
 121        for (usage = 0; nodes && pages; pages -= nodes->size, nodes++)
 122                usage += amdgpu_vram_mgr_vis_size(adev, nodes);
 123
 124        return usage;
 125}
 126
 127/**
 128 * amdgpu_vram_mgr_new - allocate new ranges
 129 *
 130 * @man: TTM memory type manager
 131 * @tbo: TTM BO we need this range for
 132 * @place: placement flags and restrictions
 133 * @mem: the resulting mem object
 134 *
 135 * Allocate VRAM for the given BO.
 136 */
 137static int amdgpu_vram_mgr_new(struct ttm_mem_type_manager *man,
 138                               struct ttm_buffer_object *tbo,
 139                               const struct ttm_place *place,
 140                               struct ttm_mem_reg *mem)
 141{
 142        struct amdgpu_device *adev = amdgpu_ttm_adev(man->bdev);
 143        struct amdgpu_vram_mgr *mgr = man->priv;
 144        struct drm_mm *mm = &mgr->mm;
 145        struct drm_mm_node *nodes;
 146        enum drm_mm_insert_mode mode;
 147        unsigned long lpfn, num_nodes, pages_per_node, pages_left;
 148        uint64_t usage = 0, vis_usage = 0;
 149        unsigned i;
 150        int r;
 151
 152        lpfn = place->lpfn;
 153        if (!lpfn)
 154                lpfn = man->size;
 155
 156        if (place->flags & TTM_PL_FLAG_CONTIGUOUS ||
 157            amdgpu_vram_page_split == -1) {
 158                pages_per_node = ~0ul;
 159                num_nodes = 1;
 160        } else {
 161                pages_per_node = max((uint32_t)amdgpu_vram_page_split,
 162                                     mem->page_alignment);
 163                num_nodes = DIV_ROUND_UP(mem->num_pages, pages_per_node);
 164        }
 165
 166        nodes = kvmalloc_array(num_nodes, sizeof(*nodes),
 167                               GFP_KERNEL | __GFP_ZERO);
 168        if (!nodes)
 169                return -ENOMEM;
 170
 171        mode = DRM_MM_INSERT_BEST;
 172        if (place->flags & TTM_PL_FLAG_TOPDOWN)
 173                mode = DRM_MM_INSERT_HIGH;
 174
 175        mem->start = 0;
 176        pages_left = mem->num_pages;
 177
 178        spin_lock(&mgr->lock);
 179        for (i = 0; i < num_nodes; ++i) {
 180                unsigned long pages = min(pages_left, pages_per_node);
 181                uint32_t alignment = mem->page_alignment;
 182                unsigned long start;
 183
 184                if (pages == pages_per_node)
 185                        alignment = pages_per_node;
 186
 187                r = drm_mm_insert_node_in_range(mm, &nodes[i],
 188                                                pages, alignment, 0,
 189                                                place->fpfn, lpfn,
 190                                                mode);
 191                if (unlikely(r))
 192                        goto error;
 193
 194                usage += nodes[i].size << PAGE_SHIFT;
 195                vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]);
 196
 197                /* Calculate a virtual BO start address to easily check if
 198                 * everything is CPU accessible.
 199                 */
 200                start = nodes[i].start + nodes[i].size;
 201                if (start > mem->num_pages)
 202                        start -= mem->num_pages;
 203                else
 204                        start = 0;
 205                mem->start = max(mem->start, start);
 206                pages_left -= pages;
 207        }
 208        spin_unlock(&mgr->lock);
 209
 210        atomic64_add(usage, &mgr->usage);
 211        atomic64_add(vis_usage, &mgr->vis_usage);
 212
 213        mem->mm_node = nodes;
 214
 215        return 0;
 216
 217error:
 218        while (i--)
 219                drm_mm_remove_node(&nodes[i]);
 220        spin_unlock(&mgr->lock);
 221
 222        kvfree(nodes);
 223        return r == -ENOSPC ? 0 : r;
 224}
 225
 226/**
 227 * amdgpu_vram_mgr_del - free ranges
 228 *
 229 * @man: TTM memory type manager
 230 * @tbo: TTM BO we need this range for
 231 * @place: placement flags and restrictions
 232 * @mem: TTM memory object
 233 *
 234 * Free the allocated VRAM again.
 235 */
 236static void amdgpu_vram_mgr_del(struct ttm_mem_type_manager *man,
 237                                struct ttm_mem_reg *mem)
 238{
 239        struct amdgpu_device *adev = amdgpu_ttm_adev(man->bdev);
 240        struct amdgpu_vram_mgr *mgr = man->priv;
 241        struct drm_mm_node *nodes = mem->mm_node;
 242        uint64_t usage = 0, vis_usage = 0;
 243        unsigned pages = mem->num_pages;
 244
 245        if (!mem->mm_node)
 246                return;
 247
 248        spin_lock(&mgr->lock);
 249        while (pages) {
 250                pages -= nodes->size;
 251                drm_mm_remove_node(nodes);
 252                usage += nodes->size << PAGE_SHIFT;
 253                vis_usage += amdgpu_vram_mgr_vis_size(adev, nodes);
 254                ++nodes;
 255        }
 256        spin_unlock(&mgr->lock);
 257
 258        atomic64_sub(usage, &mgr->usage);
 259        atomic64_sub(vis_usage, &mgr->vis_usage);
 260
 261        kvfree(mem->mm_node);
 262        mem->mm_node = NULL;
 263}
 264
 265/**
 266 * amdgpu_vram_mgr_usage - how many bytes are used in this domain
 267 *
 268 * @man: TTM memory type manager
 269 *
 270 * Returns how many bytes are used in this domain.
 271 */
 272uint64_t amdgpu_vram_mgr_usage(struct ttm_mem_type_manager *man)
 273{
 274        struct amdgpu_vram_mgr *mgr = man->priv;
 275
 276        return atomic64_read(&mgr->usage);
 277}
 278
 279/**
 280 * amdgpu_vram_mgr_vis_usage - how many bytes are used in the visible part
 281 *
 282 * @man: TTM memory type manager
 283 *
 284 * Returns how many bytes are used in the visible part of VRAM
 285 */
 286uint64_t amdgpu_vram_mgr_vis_usage(struct ttm_mem_type_manager *man)
 287{
 288        struct amdgpu_vram_mgr *mgr = man->priv;
 289
 290        return atomic64_read(&mgr->vis_usage);
 291}
 292
 293/**
 294 * amdgpu_vram_mgr_debug - dump VRAM table
 295 *
 296 * @man: TTM memory type manager
 297 * @printer: DRM printer to use
 298 *
 299 * Dump the table content using printk.
 300 */
 301static void amdgpu_vram_mgr_debug(struct ttm_mem_type_manager *man,
 302                                  struct drm_printer *printer)
 303{
 304        struct amdgpu_vram_mgr *mgr = man->priv;
 305
 306        spin_lock(&mgr->lock);
 307        drm_mm_print(&mgr->mm, printer);
 308        spin_unlock(&mgr->lock);
 309
 310        drm_printf(printer, "man size:%llu pages, ram usage:%lluMB, vis usage:%lluMB\n",
 311                   man->size, amdgpu_vram_mgr_usage(man) >> 20,
 312                   amdgpu_vram_mgr_vis_usage(man) >> 20);
 313}
 314
 315const struct ttm_mem_type_manager_func amdgpu_vram_mgr_func = {
 316        .init           = amdgpu_vram_mgr_init,
 317        .takedown       = amdgpu_vram_mgr_fini,
 318        .get_node       = amdgpu_vram_mgr_new,
 319        .put_node       = amdgpu_vram_mgr_del,
 320        .debug          = amdgpu_vram_mgr_debug
 321};
 322