}
}
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-static void drm_bo_delayed_workqueue(void *data)
-#else
static void drm_bo_delayed_workqueue(struct work_struct *work)
-#endif
{
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
- struct drm_device *dev = (struct drm_device *) data;
- struct drm_buffer_manager *bm = &dev->bm;
-#else
struct drm_buffer_manager *bm =
container_of(work, struct drm_buffer_manager, wq.work);
struct drm_device *dev = container_of(bm, struct drm_device, bm);
-#endif
DRM_DEBUG("Delayed delete Worker\n");
if (ret)
goto out_unlock;
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
- INIT_WORK(&bm->wq, &drm_bo_delayed_workqueue, dev);
-#else
INIT_DELAYED_WORK(&bm->wq, drm_bo_delayed_workqueue);
-#endif
bm->initialized = 1;
bm->nice_mode = 1;
atomic_set(&bm->count, 0);
struct drm_mem_type_manager man[DRM_BO_MEM_TYPES];
struct list_head unfenced;
struct list_head ddestroy;
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
- struct work_struct wq;
-#else
struct delayed_work wq;
-#endif
uint32_t fence_type;
unsigned long cur_pages;
atomic_t count;
static void
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
-via_dmablit_workqueue(void *data)
-#else
via_dmablit_workqueue(struct work_struct *work)
-#endif
{
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
- drm_via_blitq_t *blitq = (drm_via_blitq_t *) data;
-#else
drm_via_blitq_t *blitq = container_of(work, drm_via_blitq_t, wq);
-#endif
struct drm_device *dev = blitq->dev;
unsigned long irqsave;
drm_via_sg_info_t *cur_sg;
DRM_INIT_WAITQUEUE(blitq->blit_queue + j);
}
DRM_INIT_WAITQUEUE(&blitq->busy_queue);
-#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,20)
- INIT_WORK(&blitq->wq, via_dmablit_workqueue, blitq);
-#else
INIT_WORK(&blitq->wq, via_dmablit_workqueue);
-#endif
init_timer(&blitq->poll_timer);
blitq->poll_timer.function = &via_dmablit_timer;
blitq->poll_timer.data = (unsigned long) blitq;