drm: Drop superfluous "select VT_HW_CONSOLE_BINDING"
[linux-2.6-block.git] / drivers / gpu / drm / cirrus / cirrus_fbdev.c
CommitLineData
f9aa76a8
DA
1/*
2 * Copyright 2012 Red Hat
3 *
4 * This file is subject to the terms and conditions of the GNU General
5 * Public License version 2. See the file COPYING in the main
6 * directory of this archive for more details.
7 *
8 * Authors: Matthew Garrett
9 * Dave Airlie
10 */
11#include <linux/module.h>
760285e7
DH
12#include <drm/drmP.h>
13#include <drm/drm_fb_helper.h>
76a39dbf 14#include <drm/drm_crtc_helper.h>
f9aa76a8
DA
15
16#include <linux/fb.h>
17
18#include "cirrus_drv.h"
19
20static void cirrus_dirty_update(struct cirrus_fbdev *afbdev,
21 int x, int y, int width, int height)
22{
23 int i;
24 struct drm_gem_object *obj;
25 struct cirrus_bo *bo;
26 int src_offset, dst_offset;
27 int bpp = (afbdev->gfb.base.bits_per_pixel + 7)/8;
19d4b72c 28 int ret = -EBUSY;
f9aa76a8 29 bool unmap = false;
f3b2bbdc
DA
30 bool store_for_later = false;
31 int x2, y2;
32 unsigned long flags;
f9aa76a8
DA
33
34 obj = afbdev->gfb.obj;
35 bo = gem_to_cirrus_bo(obj);
36
f3b2bbdc
DA
37 /*
38 * try and reserve the BO, if we fail with busy
39 * then the BO is being moved and we should
40 * store up the damage until later.
41 */
8b7ad1bb 42 if (drm_can_sleep())
19d4b72c 43 ret = cirrus_bo_reserve(bo, true);
f9aa76a8 44 if (ret) {
f3b2bbdc
DA
45 if (ret != -EBUSY)
46 return;
47 store_for_later = true;
48 }
49
50 x2 = x + width - 1;
51 y2 = y + height - 1;
52 spin_lock_irqsave(&afbdev->dirty_lock, flags);
53
54 if (afbdev->y1 < y)
55 y = afbdev->y1;
56 if (afbdev->y2 > y2)
57 y2 = afbdev->y2;
58 if (afbdev->x1 < x)
59 x = afbdev->x1;
60 if (afbdev->x2 > x2)
61 x2 = afbdev->x2;
62
63 if (store_for_later) {
64 afbdev->x1 = x;
65 afbdev->x2 = x2;
66 afbdev->y1 = y;
67 afbdev->y2 = y2;
68 spin_unlock_irqrestore(&afbdev->dirty_lock, flags);
f9aa76a8
DA
69 return;
70 }
71
f3b2bbdc
DA
72 afbdev->x1 = afbdev->y1 = INT_MAX;
73 afbdev->x2 = afbdev->y2 = 0;
74 spin_unlock_irqrestore(&afbdev->dirty_lock, flags);
75
f9aa76a8
DA
76 if (!bo->kmap.virtual) {
77 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap);
78 if (ret) {
79 DRM_ERROR("failed to kmap fb updates\n");
80 cirrus_bo_unreserve(bo);
81 return;
82 }
83 unmap = true;
84 }
85 for (i = y; i < y + height; i++) {
86 /* assume equal stride for now */
87 src_offset = dst_offset = i * afbdev->gfb.base.pitches[0] + (x * bpp);
88 memcpy_toio(bo->kmap.virtual + src_offset, afbdev->sysram + src_offset, width * bpp);
89
90 }
91 if (unmap)
92 ttm_bo_kunmap(&bo->kmap);
93
94 cirrus_bo_unreserve(bo);
95}
96
97static void cirrus_fillrect(struct fb_info *info,
98 const struct fb_fillrect *rect)
99{
100 struct cirrus_fbdev *afbdev = info->par;
101 sys_fillrect(info, rect);
102 cirrus_dirty_update(afbdev, rect->dx, rect->dy, rect->width,
103 rect->height);
104}
105
106static void cirrus_copyarea(struct fb_info *info,
107 const struct fb_copyarea *area)
108{
109 struct cirrus_fbdev *afbdev = info->par;
110 sys_copyarea(info, area);
111 cirrus_dirty_update(afbdev, area->dx, area->dy, area->width,
112 area->height);
113}
114
115static void cirrus_imageblit(struct fb_info *info,
116 const struct fb_image *image)
117{
118 struct cirrus_fbdev *afbdev = info->par;
119 sys_imageblit(info, image);
120 cirrus_dirty_update(afbdev, image->dx, image->dy, image->width,
121 image->height);
122}
123
124
125static struct fb_ops cirrusfb_ops = {
126 .owner = THIS_MODULE,
127 .fb_check_var = drm_fb_helper_check_var,
128 .fb_set_par = drm_fb_helper_set_par,
129 .fb_fillrect = cirrus_fillrect,
130 .fb_copyarea = cirrus_copyarea,
131 .fb_imageblit = cirrus_imageblit,
132 .fb_pan_display = drm_fb_helper_pan_display,
133 .fb_blank = drm_fb_helper_blank,
134 .fb_setcmap = drm_fb_helper_setcmap,
135};
136
137static int cirrusfb_create_object(struct cirrus_fbdev *afbdev,
138 struct drm_mode_fb_cmd2 *mode_cmd,
139 struct drm_gem_object **gobj_p)
140{
141 struct drm_device *dev = afbdev->helper.dev;
8975626e 142 struct cirrus_device *cdev = dev->dev_private;
f9aa76a8
DA
143 u32 bpp, depth;
144 u32 size;
145 struct drm_gem_object *gobj;
146
147 int ret = 0;
148 drm_fb_get_bpp_depth(mode_cmd->pixel_format, &depth, &bpp);
149
8975626e
ZR
150 if (!cirrus_check_framebuffer(cdev, mode_cmd->width, mode_cmd->height,
151 bpp, mode_cmd->pitches[0]))
f9aa76a8 152 return -EINVAL;
8975626e 153
f9aa76a8
DA
154 size = mode_cmd->pitches[0] * mode_cmd->height;
155 ret = cirrus_gem_create(dev, size, true, &gobj);
156 if (ret)
157 return ret;
158
159 *gobj_p = gobj;
160 return ret;
161}
162
cd5428a5 163static int cirrusfb_create(struct drm_fb_helper *helper,
f9aa76a8
DA
164 struct drm_fb_helper_surface_size *sizes)
165{
ea0622cf
FF
166 struct cirrus_fbdev *gfbdev =
167 container_of(helper, struct cirrus_fbdev, helper);
f9aa76a8
DA
168 struct drm_device *dev = gfbdev->helper.dev;
169 struct cirrus_device *cdev = gfbdev->helper.dev->dev_private;
170 struct fb_info *info;
171 struct drm_framebuffer *fb;
172 struct drm_mode_fb_cmd2 mode_cmd;
173 struct device *device = &dev->pdev->dev;
174 void *sysram;
175 struct drm_gem_object *gobj = NULL;
176 struct cirrus_bo *bo = NULL;
177 int size, ret;
178
179 mode_cmd.width = sizes->surface_width;
180 mode_cmd.height = sizes->surface_height;
181 mode_cmd.pitches[0] = mode_cmd.width * ((sizes->surface_bpp + 7) / 8);
182 mode_cmd.pixel_format = drm_mode_legacy_fb_format(sizes->surface_bpp,
183 sizes->surface_depth);
184 size = mode_cmd.pitches[0] * mode_cmd.height;
185
186 ret = cirrusfb_create_object(gfbdev, &mode_cmd, &gobj);
187 if (ret) {
188 DRM_ERROR("failed to create fbcon backing object %d\n", ret);
189 return ret;
190 }
191
192 bo = gem_to_cirrus_bo(gobj);
193
194 sysram = vmalloc(size);
195 if (!sysram)
196 return -ENOMEM;
197
198 info = framebuffer_alloc(0, device);
199 if (info == NULL)
200 return -ENOMEM;
201
202 info->par = gfbdev;
203
204 ret = cirrus_framebuffer_init(cdev->dev, &gfbdev->gfb, &mode_cmd, gobj);
205 if (ret)
206 return ret;
207
208 gfbdev->sysram = sysram;
209 gfbdev->size = size;
210
211 fb = &gfbdev->gfb.base;
212 if (!fb) {
213 DRM_INFO("fb is NULL\n");
214 return -EINVAL;
215 }
216
217 /* setup helper */
218 gfbdev->helper.fb = fb;
219 gfbdev->helper.fbdev = info;
220
221 strcpy(info->fix.id, "cirrusdrmfb");
222
223
224 info->flags = FBINFO_DEFAULT;
225 info->fbops = &cirrusfb_ops;
226
227 drm_fb_helper_fill_fix(info, fb->pitches[0], fb->depth);
228 drm_fb_helper_fill_var(info, &gfbdev->helper, sizes->fb_width,
229 sizes->fb_height);
230
231 /* setup aperture base/size for vesafb takeover */
232 info->apertures = alloc_apertures(1);
233 if (!info->apertures) {
234 ret = -ENOMEM;
235 goto out_iounmap;
236 }
237 info->apertures->ranges[0].base = cdev->dev->mode_config.fb_base;
238 info->apertures->ranges[0].size = cdev->mc.vram_size;
239
99d4a8ae
MK
240 info->fix.smem_start = cdev->dev->mode_config.fb_base;
241 info->fix.smem_len = cdev->mc.vram_size;
242
f9aa76a8
DA
243 info->screen_base = sysram;
244 info->screen_size = size;
245
246 info->fix.mmio_start = 0;
247 info->fix.mmio_len = 0;
248
249 ret = fb_alloc_cmap(&info->cmap, 256, 0);
250 if (ret) {
251 DRM_ERROR("%s: can't allocate color map\n", info->fix.id);
252 ret = -ENOMEM;
253 goto out_iounmap;
254 }
255
256 DRM_INFO("fb mappable at 0x%lX\n", info->fix.smem_start);
257 DRM_INFO("vram aper at 0x%lX\n", (unsigned long)info->fix.smem_start);
258 DRM_INFO("size %lu\n", (unsigned long)info->fix.smem_len);
259 DRM_INFO("fb depth is %d\n", fb->depth);
260 DRM_INFO(" pitch is %d\n", fb->pitches[0]);
261
262 return 0;
263out_iounmap:
264 return ret;
265}
266
f9aa76a8
DA
267static int cirrus_fbdev_destroy(struct drm_device *dev,
268 struct cirrus_fbdev *gfbdev)
269{
270 struct fb_info *info;
271 struct cirrus_framebuffer *gfb = &gfbdev->gfb;
272
273 if (gfbdev->helper.fbdev) {
274 info = gfbdev->helper.fbdev;
275
276 unregister_framebuffer(info);
277 if (info->cmap.len)
278 fb_dealloc_cmap(&info->cmap);
279 framebuffer_release(info);
280 }
281
282 if (gfb->obj) {
283 drm_gem_object_unreference_unlocked(gfb->obj);
284 gfb->obj = NULL;
285 }
286
287 vfree(gfbdev->sysram);
288 drm_fb_helper_fini(&gfbdev->helper);
36206361 289 drm_framebuffer_unregister_private(&gfb->base);
f9aa76a8
DA
290 drm_framebuffer_cleanup(&gfb->base);
291
292 return 0;
293}
294
3a493879 295static const struct drm_fb_helper_funcs cirrus_fb_helper_funcs = {
f9aa76a8
DA
296 .gamma_set = cirrus_crtc_fb_gamma_set,
297 .gamma_get = cirrus_crtc_fb_gamma_get,
cd5428a5 298 .fb_probe = cirrusfb_create,
f9aa76a8
DA
299};
300
301int cirrus_fbdev_init(struct cirrus_device *cdev)
302{
303 struct cirrus_fbdev *gfbdev;
304 int ret;
305 int bpp_sel = 24;
306
307 /*bpp_sel = 8;*/
308 gfbdev = kzalloc(sizeof(struct cirrus_fbdev), GFP_KERNEL);
309 if (!gfbdev)
310 return -ENOMEM;
311
312 cdev->mode_info.gfbdev = gfbdev;
f3b2bbdc 313 spin_lock_init(&gfbdev->dirty_lock);
f9aa76a8 314
10a23102
TR
315 drm_fb_helper_prepare(cdev->dev, &gfbdev->helper,
316 &cirrus_fb_helper_funcs);
317
f9aa76a8
DA
318 ret = drm_fb_helper_init(cdev->dev, &gfbdev->helper,
319 cdev->num_crtc, CIRRUSFB_CONN_LIMIT);
320 if (ret) {
321 kfree(gfbdev);
322 return ret;
323 }
324 drm_fb_helper_single_add_all_connectors(&gfbdev->helper);
76a39dbf
DV
325
326 /* disable all the possible outputs/crtcs before entering KMS mode */
327 drm_helper_disable_unused_functions(cdev->dev);
f9aa76a8
DA
328 drm_fb_helper_initial_config(&gfbdev->helper, bpp_sel);
329
330 return 0;
331}
332
333void cirrus_fbdev_fini(struct cirrus_device *cdev)
334{
335 if (!cdev->mode_info.gfbdev)
336 return;
337
338 cirrus_fbdev_destroy(cdev->dev, cdev->mode_info.gfbdev);
339 kfree(cdev->mode_info.gfbdev);
340 cdev->mode_info.gfbdev = NULL;
341}