i915: cleanup most of the whitespace
parent
427cc561b3
commit
9493ce6ca3
|
@ -33,14 +33,14 @@
|
|||
#include "i915_drm.h"
|
||||
#include "i915_drv.h"
|
||||
|
||||
struct drm_ttm_backend *i915_create_ttm_backend_entry(struct drm_device * dev)
|
||||
struct drm_ttm_backend *i915_create_ttm_backend_entry(struct drm_device *dev)
|
||||
{
|
||||
return drm_agp_init_ttm(dev);
|
||||
}
|
||||
|
||||
int i915_fence_types(struct drm_buffer_object *bo,
|
||||
uint32_t * fclass,
|
||||
uint32_t * type)
|
||||
uint32_t *fclass,
|
||||
uint32_t *type)
|
||||
{
|
||||
if (bo->mem.mask & (DRM_BO_FLAG_READ | DRM_BO_FLAG_WRITE))
|
||||
*type = 3;
|
||||
|
@ -49,7 +49,7 @@ int i915_fence_types(struct drm_buffer_object *bo,
|
|||
return 0;
|
||||
}
|
||||
|
||||
int i915_invalidate_caches(struct drm_device * dev, uint64_t flags)
|
||||
int i915_invalidate_caches(struct drm_device *dev, uint64_t flags)
|
||||
{
|
||||
/*
|
||||
* FIXME: Only emit once per batchbuffer submission.
|
||||
|
@ -65,8 +65,8 @@ int i915_invalidate_caches(struct drm_device * dev, uint64_t flags)
|
|||
return i915_emit_mi_flush(dev, flush_cmd);
|
||||
}
|
||||
|
||||
int i915_init_mem_type(struct drm_device * dev, uint32_t type,
|
||||
struct drm_mem_type_manager * man)
|
||||
int i915_init_mem_type(struct drm_device *dev, uint32_t type,
|
||||
struct drm_mem_type_manager *man)
|
||||
{
|
||||
switch (type) {
|
||||
case DRM_BO_MEM_LOCAL:
|
||||
|
@ -226,25 +226,24 @@ out_cleanup:
|
|||
#endif
|
||||
|
||||
/*
|
||||
* Disable i915_move_flip for now, since we can't guarantee that the hardware lock
|
||||
* is held here. To re-enable we need to make sure either
|
||||
* Disable i915_move_flip for now, since we can't guarantee that the hardware
|
||||
* lock is held here. To re-enable we need to make sure either
|
||||
* a) The X server is using DRM to submit commands to the ring, or
|
||||
* b) DRM can use the HP ring for these blits. This means i915 needs to implement
|
||||
* a new ring submission mechanism and fence class.
|
||||
* b) DRM can use the HP ring for these blits. This means i915 needs to
|
||||
* implement a new ring submission mechanism and fence class.
|
||||
*/
|
||||
|
||||
int i915_move(struct drm_buffer_object * bo,
|
||||
int evict, int no_wait, struct drm_bo_mem_reg * new_mem)
|
||||
int i915_move(struct drm_buffer_object *bo,
|
||||
int evict, int no_wait, struct drm_bo_mem_reg *new_mem)
|
||||
{
|
||||
struct drm_bo_mem_reg *old_mem = &bo->mem;
|
||||
|
||||
if (old_mem->mem_type == DRM_BO_MEM_LOCAL) {
|
||||
return drm_bo_move_memcpy(bo, evict, no_wait, new_mem);
|
||||
} else if (new_mem->mem_type == DRM_BO_MEM_LOCAL) {
|
||||
if (0 /*i915_move_flip(bo, evict, no_wait, new_mem)*/)
|
||||
if (0) /*i915_move_flip(bo, evict, no_wait, new_mem)*/
|
||||
return drm_bo_move_memcpy(bo, evict, no_wait, new_mem);
|
||||
} else {
|
||||
if (0 /*i915_move_blit(bo, evict, no_wait, new_mem)*/)
|
||||
if (0) /*i915_move_blit(bo, evict, no_wait, new_mem)*/
|
||||
return drm_bo_move_memcpy(bo, evict, no_wait, new_mem);
|
||||
}
|
||||
return 0;
|
||||
|
@ -259,7 +258,7 @@ static inline void clflush(volatile void *__p)
|
|||
|
||||
static inline void drm_cache_flush_addr(void *virt)
|
||||
{
|
||||
int i;
|
||||
int i;
|
||||
|
||||
for (i = 0; i < PAGE_SIZE; i += boot_cpu_data.x86_clflush_size)
|
||||
clflush(virt+i);
|
||||
|
|
|
@ -38,7 +38,7 @@
|
|||
* Implements an intel sync flush operation.
|
||||
*/
|
||||
|
||||
static void i915_perform_flush(struct drm_device * dev)
|
||||
static void i915_perform_flush(struct drm_device *dev)
|
||||
{
|
||||
drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
|
||||
struct drm_fence_manager *fm = &dev->fm;
|
||||
|
@ -63,7 +63,7 @@ static void i915_perform_flush(struct drm_device * dev)
|
|||
|
||||
diff = (sequence - fc->last_exe_flush) & BREADCRUMB_MASK;
|
||||
if (diff < driver->wrap_diff && diff != 0) {
|
||||
drm_fence_handler(dev, 0, sequence,
|
||||
drm_fence_handler(dev, 0, sequence,
|
||||
DRM_FENCE_TYPE_EXE, 0);
|
||||
}
|
||||
|
||||
|
@ -110,7 +110,7 @@ static void i915_perform_flush(struct drm_device * dev)
|
|||
|
||||
}
|
||||
|
||||
void i915_poke_flush(struct drm_device * dev, uint32_t class)
|
||||
void i915_poke_flush(struct drm_device *dev, uint32_t class)
|
||||
{
|
||||
struct drm_fence_manager *fm = &dev->fm;
|
||||
unsigned long flags;
|
||||
|
@ -120,8 +120,9 @@ void i915_poke_flush(struct drm_device * dev, uint32_t class)
|
|||
write_unlock_irqrestore(&fm->lock, flags);
|
||||
}
|
||||
|
||||
int i915_fence_emit_sequence(struct drm_device * dev, uint32_t class, uint32_t flags,
|
||||
uint32_t * sequence, uint32_t * native_type)
|
||||
int i915_fence_emit_sequence(struct drm_device *dev, uint32_t class,
|
||||
uint32_t flags, uint32_t *sequence,
|
||||
uint32_t *native_type)
|
||||
{
|
||||
drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
|
||||
if (!dev_priv)
|
||||
|
@ -136,7 +137,7 @@ int i915_fence_emit_sequence(struct drm_device * dev, uint32_t class, uint32_t f
|
|||
return 0;
|
||||
}
|
||||
|
||||
void i915_fence_handler(struct drm_device * dev)
|
||||
void i915_fence_handler(struct drm_device *dev)
|
||||
{
|
||||
struct drm_fence_manager *fm = &dev->fm;
|
||||
|
||||
|
|
|
@ -267,7 +267,7 @@ extern int i915_vblank_pipe_set(struct drm_device *dev, void *data,
|
|||
struct drm_file *file_priv);
|
||||
extern int i915_vblank_pipe_get(struct drm_device *dev, void *data,
|
||||
struct drm_file *file_priv);
|
||||
extern int i915_emit_irq(struct drm_device * dev);
|
||||
extern int i915_emit_irq(struct drm_device *dev);
|
||||
extern void i915_user_irq_on(drm_i915_private_t *dev_priv);
|
||||
extern void i915_user_irq_off(drm_i915_private_t *dev_priv);
|
||||
extern int i915_vblank_swap(struct drm_device *dev, void *data,
|
||||
|
@ -303,7 +303,7 @@ extern int i915_fence_has_irq(struct drm_device *dev, uint32_t class, uint32_t f
|
|||
/* i915_buffer.c */
|
||||
extern struct drm_ttm_backend *i915_create_ttm_backend_entry(struct drm_device *dev);
|
||||
extern int i915_fence_types(struct drm_buffer_object *bo, uint32_t *fclass,
|
||||
uint32_t *type);
|
||||
uint32_t *type);
|
||||
extern int i915_invalidate_caches(struct drm_device *dev, uint64_t buffer_flags);
|
||||
extern int i915_init_mem_type(struct drm_device *dev, uint32_t type,
|
||||
struct drm_mem_type_manager *man);
|
||||
|
|
|
@ -365,9 +365,8 @@ irqreturn_t i915_driver_irq_handler(DRM_IRQ_ARGS)
|
|||
return IRQ_HANDLED;
|
||||
}
|
||||
|
||||
int i915_emit_irq(struct drm_device * dev)
|
||||
int i915_emit_irq(struct drm_device *dev)
|
||||
{
|
||||
|
||||
drm_i915_private_t *dev_priv = dev->dev_private;
|
||||
RING_LOCALS;
|
||||
|
||||
|
@ -383,8 +382,6 @@ int i915_emit_irq(struct drm_device * dev)
|
|||
ADVANCE_LP_RING();
|
||||
|
||||
return dev_priv->counter;
|
||||
|
||||
|
||||
}
|
||||
|
||||
void i915_user_irq_on(drm_i915_private_t *dev_priv)
|
||||
|
@ -471,7 +468,8 @@ int i915_driver_vblank_wait2(struct drm_device *dev, unsigned int *sequence)
|
|||
|
||||
/* Needs the lock as it touches the ring.
|
||||
*/
|
||||
int i915_irq_emit(struct drm_device *dev, void *data, struct drm_file *file_priv)
|
||||
int i915_irq_emit(struct drm_device *dev, void *data,
|
||||
struct drm_file *file_priv)
|
||||
{
|
||||
drm_i915_private_t *dev_priv = dev->dev_private;
|
||||
drm_i915_irq_emit_t *emit = data;
|
||||
|
@ -747,13 +745,14 @@ void i915_driver_irq_postinstall(struct drm_device * dev)
|
|||
* Initialize the hardware status page IRQ location.
|
||||
*/
|
||||
|
||||
I915_WRITE(I915REG_INSTPM, ( 1 << 5) | ( 1 << 21));
|
||||
I915_WRITE(I915REG_INSTPM, (1 << 5) | (1 << 21));
|
||||
}
|
||||
|
||||
void i915_driver_irq_uninstall(struct drm_device * dev)
|
||||
{
|
||||
drm_i915_private_t *dev_priv = (drm_i915_private_t *) dev->dev_private;
|
||||
u16 temp;
|
||||
|
||||
if (!dev_priv)
|
||||
return;
|
||||
|
||||
|
|
Loading…
Reference in New Issue