linux/drivers/gpu/drm/udl/udl_fb.c
<<
>>
Prefs
   1/*
   2 * Copyright (C) 2012 Red Hat
   3 *
   4 * based in parts on udlfb.c:
   5 * Copyright (C) 2009 Roberto De Ioris <roberto@unbit.it>
   6 * Copyright (C) 2009 Jaya Kumar <jayakumar.lkml@gmail.com>
   7 * Copyright (C) 2009 Bernie Thompson <bernie@plugable.com>
   8 *
   9 * This file is subject to the terms and conditions of the GNU General Public
  10 * License v2. See the file COPYING in the main directory of this archive for
  11 * more details.
  12 */
  13#include <linux/module.h>
  14#include <linux/slab.h>
  15#include <linux/fb.h>
  16#include <linux/dma-buf.h>
  17
  18#include <drm/drmP.h>
  19#include <drm/drm_crtc.h>
  20#include <drm/drm_crtc_helper.h>
  21#include "udl_drv.h"
  22
  23#include <drm/drm_fb_helper.h>
  24
  25#define DL_DEFIO_WRITE_DELAY    (HZ/20) /* fb_deferred_io.delay in jiffies */
  26
  27static int fb_defio = 0;  /* Optionally enable experimental fb_defio mmap support */
  28static int fb_bpp = 16;
  29
  30module_param(fb_bpp, int, S_IWUSR | S_IRUSR | S_IWGRP | S_IRGRP);
  31module_param(fb_defio, int, S_IWUSR | S_IRUSR | S_IWGRP | S_IRGRP);
  32
  33struct udl_fbdev {
  34        struct drm_fb_helper helper;
  35        struct udl_framebuffer ufb;
  36        struct list_head fbdev_list;
  37        int fb_count;
  38};
  39
  40#define DL_ALIGN_UP(x, a) ALIGN(x, a)
  41#define DL_ALIGN_DOWN(x, a) ALIGN(x-(a-1), a)
  42
  43/** Read the red component (0..255) of a 32 bpp colour. */
  44#define DLO_RGB_GETRED(col) (uint8_t)((col) & 0xFF)
  45
  46/** Read the green component (0..255) of a 32 bpp colour. */
  47#define DLO_RGB_GETGRN(col) (uint8_t)(((col) >> 8) & 0xFF)
  48
  49/** Read the blue component (0..255) of a 32 bpp colour. */
  50#define DLO_RGB_GETBLU(col) (uint8_t)(((col) >> 16) & 0xFF)
  51
  52/** Return red/green component of a 16 bpp colour number. */
  53#define DLO_RG16(red, grn) (uint8_t)((((red) & 0xF8) | ((grn) >> 5)) & 0xFF)
  54
  55/** Return green/blue component of a 16 bpp colour number. */
  56#define DLO_GB16(grn, blu) (uint8_t)(((((grn) & 0x1C) << 3) | ((blu) >> 3)) & 0xFF)
  57
  58/** Return 8 bpp colour number from red, green and blue components. */
  59#define DLO_RGB8(red, grn, blu) ((((red) << 5) | (((grn) & 3) << 3) | ((blu) & 7)) & 0xFF)
  60
  61#if 0
  62static uint8_t rgb8(uint32_t col)
  63{
  64        uint8_t red = DLO_RGB_GETRED(col);
  65        uint8_t grn = DLO_RGB_GETGRN(col);
  66        uint8_t blu = DLO_RGB_GETBLU(col);
  67
  68        return DLO_RGB8(red, grn, blu);
  69}
  70
  71static uint16_t rgb16(uint32_t col)
  72{
  73        uint8_t red = DLO_RGB_GETRED(col);
  74        uint8_t grn = DLO_RGB_GETGRN(col);
  75        uint8_t blu = DLO_RGB_GETBLU(col);
  76
  77        return (DLO_RG16(red, grn) << 8) + DLO_GB16(grn, blu);
  78}
  79#endif
  80
  81/*
  82 * NOTE: fb_defio.c is holding info->fbdefio.mutex
  83 *   Touching ANY framebuffer memory that triggers a page fault
  84 *   in fb_defio will cause a deadlock, when it also tries to
  85 *   grab the same mutex.
  86 */
  87static void udlfb_dpy_deferred_io(struct fb_info *info,
  88                                  struct list_head *pagelist)
  89{
  90        struct page *cur;
  91        struct fb_deferred_io *fbdefio = info->fbdefio;
  92        struct udl_fbdev *ufbdev = info->par;
  93        struct drm_device *dev = ufbdev->ufb.base.dev;
  94        struct udl_device *udl = dev->dev_private;
  95        struct urb *urb;
  96        char *cmd;
  97        cycles_t start_cycles, end_cycles;
  98        int bytes_sent = 0;
  99        int bytes_identical = 0;
 100        int bytes_rendered = 0;
 101
 102        if (!fb_defio)
 103                return;
 104
 105        start_cycles = get_cycles();
 106
 107        urb = udl_get_urb(dev);
 108        if (!urb)
 109                return;
 110
 111        cmd = urb->transfer_buffer;
 112
 113        /* walk the written page list and render each to device */
 114        list_for_each_entry(cur, &fbdefio->pagelist, lru) {
 115
 116                if (udl_render_hline(dev, (ufbdev->ufb.base.bits_per_pixel / 8),
 117                                     &urb, (char *) info->fix.smem_start,
 118                                     &cmd, cur->index << PAGE_SHIFT,
 119                                     cur->index << PAGE_SHIFT,
 120                                     PAGE_SIZE, &bytes_identical, &bytes_sent))
 121                        goto error;
 122                bytes_rendered += PAGE_SIZE;
 123        }
 124
 125        if (cmd > (char *) urb->transfer_buffer) {
 126                /* Send partial buffer remaining before exiting */
 127                int len = cmd - (char *) urb->transfer_buffer;
 128                udl_submit_urb(dev, urb, len);
 129                bytes_sent += len;
 130        } else
 131                udl_urb_completion(urb);
 132
 133error:
 134        atomic_add(bytes_sent, &udl->bytes_sent);
 135        atomic_add(bytes_identical, &udl->bytes_identical);
 136        atomic_add(bytes_rendered, &udl->bytes_rendered);
 137        end_cycles = get_cycles();
 138        atomic_add(((unsigned int) ((end_cycles - start_cycles)
 139                    >> 10)), /* Kcycles */
 140                   &udl->cpu_kcycles_used);
 141}
 142
 143int udl_handle_damage(struct udl_framebuffer *fb, int x, int y,
 144                      int width, int height)
 145{
 146        struct drm_device *dev = fb->base.dev;
 147        struct udl_device *udl = dev->dev_private;
 148        int i, ret;
 149        char *cmd;
 150        cycles_t start_cycles, end_cycles;
 151        int bytes_sent = 0;
 152        int bytes_identical = 0;
 153        struct urb *urb;
 154        int aligned_x;
 155        int bpp = (fb->base.bits_per_pixel / 8);
 156        int x2, y2;
 157        bool store_for_later = false;
 158        unsigned long flags;
 159
 160        if (!fb->active_16)
 161                return 0;
 162
 163        if (!fb->obj->vmapping) {
 164                ret = udl_gem_vmap(fb->obj);
 165                if (ret == -ENOMEM) {
 166                        DRM_ERROR("failed to vmap fb\n");
 167                        return 0;
 168                }
 169                if (!fb->obj->vmapping) {
 170                        DRM_ERROR("failed to vmapping\n");
 171                        return 0;
 172                }
 173        }
 174
 175        aligned_x = DL_ALIGN_DOWN(x, sizeof(unsigned long));
 176        width = DL_ALIGN_UP(width + (x-aligned_x), sizeof(unsigned long));
 177        x = aligned_x;
 178
 179        if ((width <= 0) ||
 180            (x + width > fb->base.width) ||
 181            (y + height > fb->base.height))
 182                return -EINVAL;
 183
 184        /* if we are in atomic just store the info
 185           can't test inside spin lock */
 186        if (in_atomic())
 187                store_for_later = true;
 188
 189        x2 = x + width - 1;
 190        y2 = y + height - 1;
 191
 192        spin_lock_irqsave(&fb->dirty_lock, flags);
 193
 194        if (fb->y1 < y)
 195                y = fb->y1;
 196        if (fb->y2 > y2)
 197                y2 = fb->y2;
 198        if (fb->x1 < x)
 199                x = fb->x1;
 200        if (fb->x2 > x2)
 201                x2 = fb->x2;
 202
 203        if (store_for_later) {
 204                fb->x1 = x;
 205                fb->x2 = x2;
 206                fb->y1 = y;
 207                fb->y2 = y2;
 208                spin_unlock_irqrestore(&fb->dirty_lock, flags);
 209                return 0;
 210        }
 211
 212        fb->x1 = fb->y1 = INT_MAX;
 213        fb->x2 = fb->y2 = 0;
 214
 215        spin_unlock_irqrestore(&fb->dirty_lock, flags);
 216        start_cycles = get_cycles();
 217
 218        urb = udl_get_urb(dev);
 219        if (!urb)
 220                return 0;
 221        cmd = urb->transfer_buffer;
 222
 223        for (i = y; i <= y2 ; i++) {
 224                const int line_offset = fb->base.pitches[0] * i;
 225                const int byte_offset = line_offset + (x * bpp);
 226                const int dev_byte_offset = (fb->base.width * bpp * i) + (x * bpp);
 227                if (udl_render_hline(dev, bpp, &urb,
 228                                     (char *) fb->obj->vmapping,
 229                                     &cmd, byte_offset, dev_byte_offset,
 230                                     (x2 - x + 1) * bpp,
 231                                     &bytes_identical, &bytes_sent))
 232                        goto error;
 233        }
 234
 235        if (cmd > (char *) urb->transfer_buffer) {
 236                /* Send partial buffer remaining before exiting */
 237                int len = cmd - (char *) urb->transfer_buffer;
 238                ret = udl_submit_urb(dev, urb, len);
 239                bytes_sent += len;
 240        } else
 241                udl_urb_completion(urb);
 242
 243error:
 244        atomic_add(bytes_sent, &udl->bytes_sent);
 245        atomic_add(bytes_identical, &udl->bytes_identical);
 246        atomic_add(width*height*bpp, &udl->bytes_rendered);
 247        end_cycles = get_cycles();
 248        atomic_add(((unsigned int) ((end_cycles - start_cycles)
 249                    >> 10)), /* Kcycles */
 250                   &udl->cpu_kcycles_used);
 251
 252        return 0;
 253}
 254
 255static int udl_fb_mmap(struct fb_info *info, struct vm_area_struct *vma)
 256{
 257        unsigned long start = vma->vm_start;
 258        unsigned long size = vma->vm_end - vma->vm_start;
 259        unsigned long offset = vma->vm_pgoff << PAGE_SHIFT;
 260        unsigned long page, pos;
 261
 262        if (offset + size > info->fix.smem_len)
 263                return -EINVAL;
 264
 265        pos = (unsigned long)info->fix.smem_start + offset;
 266
 267        pr_notice("mmap() framebuffer addr:%lu size:%lu\n",
 268                  pos, size);
 269
 270        while (size > 0) {
 271                page = vmalloc_to_pfn((void *)pos);
 272                if (remap_pfn_range(vma, start, page, PAGE_SIZE, PAGE_SHARED))
 273                        return -EAGAIN;
 274
 275                start += PAGE_SIZE;
 276                pos += PAGE_SIZE;
 277                if (size > PAGE_SIZE)
 278                        size -= PAGE_SIZE;
 279                else
 280                        size = 0;
 281        }
 282
 283        /* VM_IO | VM_DONTEXPAND | VM_DONTDUMP are set by remap_pfn_range() */
 284        return 0;
 285}
 286
 287static void udl_fb_fillrect(struct fb_info *info, const struct fb_fillrect *rect)
 288{
 289        struct udl_fbdev *ufbdev = info->par;
 290
 291        sys_fillrect(info, rect);
 292
 293        udl_handle_damage(&ufbdev->ufb, rect->dx, rect->dy, rect->width,
 294                          rect->height);
 295}
 296
 297static void udl_fb_copyarea(struct fb_info *info, const struct fb_copyarea *region)
 298{
 299        struct udl_fbdev *ufbdev = info->par;
 300
 301        sys_copyarea(info, region);
 302
 303        udl_handle_damage(&ufbdev->ufb, region->dx, region->dy, region->width,
 304                          region->height);
 305}
 306
 307static void udl_fb_imageblit(struct fb_info *info, const struct fb_image *image)
 308{
 309        struct udl_fbdev *ufbdev = info->par;
 310
 311        sys_imageblit(info, image);
 312
 313        udl_handle_damage(&ufbdev->ufb, image->dx, image->dy, image->width,
 314                          image->height);
 315}
 316
 317/*
 318 * It's common for several clients to have framebuffer open simultaneously.
 319 * e.g. both fbcon and X. Makes things interesting.
 320 * Assumes caller is holding info->lock (for open and release at least)
 321 */
 322static int udl_fb_open(struct fb_info *info, int user)
 323{
 324        struct udl_fbdev *ufbdev = info->par;
 325        struct drm_device *dev = ufbdev->ufb.base.dev;
 326        struct udl_device *udl = dev->dev_private;
 327
 328        /* If the USB device is gone, we don't accept new opens */
 329        if (drm_device_is_unplugged(udl->ddev))
 330                return -ENODEV;
 331
 332        ufbdev->fb_count++;
 333
 334        if (fb_defio && (info->fbdefio == NULL)) {
 335                /* enable defio at last moment if not disabled by client */
 336
 337                struct fb_deferred_io *fbdefio;
 338
 339                fbdefio = kmalloc(sizeof(struct fb_deferred_io), GFP_KERNEL);
 340
 341                if (fbdefio) {
 342                        fbdefio->delay = DL_DEFIO_WRITE_DELAY;
 343                        fbdefio->deferred_io = udlfb_dpy_deferred_io;
 344                }
 345
 346                info->fbdefio = fbdefio;
 347                fb_deferred_io_init(info);
 348        }
 349
 350        pr_notice("open /dev/fb%d user=%d fb_info=%p count=%d\n",
 351                  info->node, user, info, ufbdev->fb_count);
 352
 353        return 0;
 354}
 355
 356
 357/*
 358 * Assumes caller is holding info->lock mutex (for open and release at least)
 359 */
 360static int udl_fb_release(struct fb_info *info, int user)
 361{
 362        struct udl_fbdev *ufbdev = info->par;
 363
 364        ufbdev->fb_count--;
 365
 366        if ((ufbdev->fb_count == 0) && (info->fbdefio)) {
 367                fb_deferred_io_cleanup(info);
 368                kfree(info->fbdefio);
 369                info->fbdefio = NULL;
 370                info->fbops->fb_mmap = udl_fb_mmap;
 371        }
 372
 373        pr_warn("released /dev/fb%d user=%d count=%d\n",
 374                info->node, user, ufbdev->fb_count);
 375
 376        return 0;
 377}
 378
 379static struct fb_ops udlfb_ops = {
 380        .owner = THIS_MODULE,
 381        .fb_check_var = drm_fb_helper_check_var,
 382        .fb_set_par = drm_fb_helper_set_par,
 383        .fb_fillrect = udl_fb_fillrect,
 384        .fb_copyarea = udl_fb_copyarea,
 385        .fb_imageblit = udl_fb_imageblit,
 386        .fb_pan_display = drm_fb_helper_pan_display,
 387        .fb_blank = drm_fb_helper_blank,
 388        .fb_setcmap = drm_fb_helper_setcmap,
 389        .fb_debug_enter = drm_fb_helper_debug_enter,
 390        .fb_debug_leave = drm_fb_helper_debug_leave,
 391        .fb_mmap = udl_fb_mmap,
 392        .fb_open = udl_fb_open,
 393        .fb_release = udl_fb_release,
 394};
 395
 396static int udl_user_framebuffer_dirty(struct drm_framebuffer *fb,
 397                                      struct drm_file *file,
 398                                      unsigned flags, unsigned color,
 399                                      struct drm_clip_rect *clips,
 400                                      unsigned num_clips)
 401{
 402        struct udl_framebuffer *ufb = to_udl_fb(fb);
 403        int i;
 404        int ret = 0;
 405
 406        if (!ufb->active_16)
 407                return 0;
 408
 409        if (ufb->obj->base.import_attach) {
 410                ret = dma_buf_begin_cpu_access(ufb->obj->base.import_attach->dmabuf,
 411                                               0, ufb->obj->base.size,
 412                                               DMA_FROM_DEVICE);
 413                if (ret)
 414                        return ret;
 415        }
 416
 417        for (i = 0; i < num_clips; i++) {
 418                ret = udl_handle_damage(ufb, clips[i].x1, clips[i].y1,
 419                                  clips[i].x2 - clips[i].x1,
 420                                  clips[i].y2 - clips[i].y1);
 421                if (ret)
 422                        break;
 423        }
 424
 425        if (ufb->obj->base.import_attach) {
 426                dma_buf_end_cpu_access(ufb->obj->base.import_attach->dmabuf,
 427                                       0, ufb->obj->base.size,
 428                                       DMA_FROM_DEVICE);
 429        }
 430        return ret;
 431}
 432
 433static void udl_user_framebuffer_destroy(struct drm_framebuffer *fb)
 434{
 435        struct udl_framebuffer *ufb = to_udl_fb(fb);
 436
 437        if (ufb->obj)
 438                drm_gem_object_unreference_unlocked(&ufb->obj->base);
 439
 440        drm_framebuffer_cleanup(fb);
 441        kfree(ufb);
 442}
 443
 444static const struct drm_framebuffer_funcs udlfb_funcs = {
 445        .destroy = udl_user_framebuffer_destroy,
 446        .dirty = udl_user_framebuffer_dirty,
 447};
 448
 449
 450static int
 451udl_framebuffer_init(struct drm_device *dev,
 452                     struct udl_framebuffer *ufb,
 453                     struct drm_mode_fb_cmd2 *mode_cmd,
 454                     struct udl_gem_object *obj)
 455{
 456        int ret;
 457
 458        spin_lock_init(&ufb->dirty_lock);
 459        ufb->obj = obj;
 460        drm_helper_mode_fill_fb_struct(&ufb->base, mode_cmd);
 461        ret = drm_framebuffer_init(dev, &ufb->base, &udlfb_funcs);
 462        return ret;
 463}
 464
 465
 466static int udlfb_create(struct drm_fb_helper *helper,
 467                        struct drm_fb_helper_surface_size *sizes)
 468{
 469        struct udl_fbdev *ufbdev = (struct udl_fbdev *)helper;
 470        struct drm_device *dev = ufbdev->helper.dev;
 471        struct fb_info *info;
 472        struct device *device = dev->dev;
 473        struct drm_framebuffer *fb;
 474        struct drm_mode_fb_cmd2 mode_cmd;
 475        struct udl_gem_object *obj;
 476        uint32_t size;
 477        int ret = 0;
 478
 479        if (sizes->surface_bpp == 24)
 480                sizes->surface_bpp = 32;
 481
 482        mode_cmd.width = sizes->surface_width;
 483        mode_cmd.height = sizes->surface_height;
 484        mode_cmd.pitches[0] = mode_cmd.width * ((sizes->surface_bpp + 7) / 8);
 485
 486        mode_cmd.pixel_format = drm_mode_legacy_fb_format(sizes->surface_bpp,
 487                                                          sizes->surface_depth);
 488
 489        size = mode_cmd.pitches[0] * mode_cmd.height;
 490        size = ALIGN(size, PAGE_SIZE);
 491
 492        obj = udl_gem_alloc_object(dev, size);
 493        if (!obj)
 494                goto out;
 495
 496        ret = udl_gem_vmap(obj);
 497        if (ret) {
 498                DRM_ERROR("failed to vmap fb\n");
 499                goto out_gfree;
 500        }
 501
 502        info = framebuffer_alloc(0, device);
 503        if (!info) {
 504                ret = -ENOMEM;
 505                goto out_gfree;
 506        }
 507        info->par = ufbdev;
 508
 509        ret = udl_framebuffer_init(dev, &ufbdev->ufb, &mode_cmd, obj);
 510        if (ret)
 511                goto out_gfree;
 512
 513        fb = &ufbdev->ufb.base;
 514
 515        ufbdev->helper.fb = fb;
 516        ufbdev->helper.fbdev = info;
 517
 518        strcpy(info->fix.id, "udldrmfb");
 519
 520        info->screen_base = ufbdev->ufb.obj->vmapping;
 521        info->fix.smem_len = size;
 522        info->fix.smem_start = (unsigned long)ufbdev->ufb.obj->vmapping;
 523
 524        info->flags = FBINFO_DEFAULT | FBINFO_CAN_FORCE_OUTPUT;
 525        info->fbops = &udlfb_ops;
 526        drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
 527        drm_fb_helper_fill_var(info, &ufbdev->helper, sizes->fb_width, sizes->fb_height);
 528
 529        ret = fb_alloc_cmap(&info->cmap, 256, 0);
 530        if (ret) {
 531                ret = -ENOMEM;
 532                goto out_gfree;
 533        }
 534
 535
 536        DRM_DEBUG_KMS("allocated %dx%d vmal %p\n",
 537                      fb->width, fb->height,
 538                      ufbdev->ufb.obj->vmapping);
 539
 540        return ret;
 541out_gfree:
 542        drm_gem_object_unreference(&ufbdev->ufb.obj->base);
 543out:
 544        return ret;
 545}
 546
 547static struct drm_fb_helper_funcs udl_fb_helper_funcs = {
 548        .fb_probe = udlfb_create,
 549};
 550
 551static void udl_fbdev_destroy(struct drm_device *dev,
 552                              struct udl_fbdev *ufbdev)
 553{
 554        struct fb_info *info;
 555        if (ufbdev->helper.fbdev) {
 556                info = ufbdev->helper.fbdev;
 557                unregister_framebuffer(info);
 558                if (info->cmap.len)
 559                        fb_dealloc_cmap(&info->cmap);
 560                framebuffer_release(info);
 561        }
 562        drm_fb_helper_fini(&ufbdev->helper);
 563        drm_framebuffer_unregister_private(&ufbdev->ufb.base);
 564        drm_framebuffer_cleanup(&ufbdev->ufb.base);
 565        drm_gem_object_unreference_unlocked(&ufbdev->ufb.obj->base);
 566}
 567
 568int udl_fbdev_init(struct drm_device *dev)
 569{
 570        struct udl_device *udl = dev->dev_private;
 571        int bpp_sel = fb_bpp;
 572        struct udl_fbdev *ufbdev;
 573        int ret;
 574
 575        ufbdev = kzalloc(sizeof(struct udl_fbdev), GFP_KERNEL);
 576        if (!ufbdev)
 577                return -ENOMEM;
 578
 579        udl->fbdev = ufbdev;
 580        ufbdev->helper.funcs = &udl_fb_helper_funcs;
 581
 582        ret = drm_fb_helper_init(dev, &ufbdev->helper,
 583                                 1, 1);
 584        if (ret) {
 585                kfree(ufbdev);
 586                return ret;
 587
 588        }
 589
 590        drm_fb_helper_single_add_all_connectors(&ufbdev->helper);
 591
 592        /* disable all the possible outputs/crtcs before entering KMS mode */
 593        drm_helper_disable_unused_functions(dev);
 594
 595        drm_fb_helper_initial_config(&ufbdev->helper, bpp_sel);
 596        return 0;
 597}
 598
 599void udl_fbdev_cleanup(struct drm_device *dev)
 600{
 601        struct udl_device *udl = dev->dev_private;
 602        if (!udl->fbdev)
 603                return;
 604
 605        udl_fbdev_destroy(dev, udl->fbdev);
 606        kfree(udl->fbdev);
 607        udl->fbdev = NULL;
 608}
 609
 610void udl_fbdev_unplug(struct drm_device *dev)
 611{
 612        struct udl_device *udl = dev->dev_private;
 613        struct udl_fbdev *ufbdev;
 614        if (!udl->fbdev)
 615                return;
 616
 617        ufbdev = udl->fbdev;
 618        if (ufbdev->helper.fbdev) {
 619                struct fb_info *info;
 620                info = ufbdev->helper.fbdev;
 621                unlink_framebuffer(info);
 622        }
 623}
 624
 625struct drm_framebuffer *
 626udl_fb_user_fb_create(struct drm_device *dev,
 627                   struct drm_file *file,
 628                   struct drm_mode_fb_cmd2 *mode_cmd)
 629{
 630        struct drm_gem_object *obj;
 631        struct udl_framebuffer *ufb;
 632        int ret;
 633        uint32_t size;
 634
 635        obj = drm_gem_object_lookup(dev, file, mode_cmd->handles[0]);
 636        if (obj == NULL)
 637                return ERR_PTR(-ENOENT);
 638
 639        size = mode_cmd->pitches[0] * mode_cmd->height;
 640        size = ALIGN(size, PAGE_SIZE);
 641
 642        if (size > obj->size) {
 643                DRM_ERROR("object size not sufficient for fb %d %zu %d %d\n", size, obj->size, mode_cmd->pitches[0], mode_cmd->height);
 644                return ERR_PTR(-ENOMEM);
 645        }
 646
 647        ufb = kzalloc(sizeof(*ufb), GFP_KERNEL);
 648        if (ufb == NULL)
 649                return ERR_PTR(-ENOMEM);
 650
 651        ret = udl_framebuffer_init(dev, ufb, mode_cmd, to_udl_bo(obj));
 652        if (ret) {
 653                kfree(ufb);
 654                return ERR_PTR(-EINVAL);
 655        }
 656        return &ufb->base;
 657}
 658