{
        bool ret;
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
        ret = _intel_set_memory_cxsr(dev_priv, enable);
        if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))
-               dev_priv->wm.vlv.cxsr = enable;
+               dev_priv->display.wm.vlv.cxsr = enable;
        else if (IS_G4X(dev_priv))
-               dev_priv->wm.g4x.cxsr = enable;
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+               dev_priv->display.wm.g4x.cxsr = enable;
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 
        return ret;
 }
 
 static int intel_wm_num_levels(struct drm_i915_private *dev_priv)
 {
-       return dev_priv->wm.max_level + 1;
+       return dev_priv->display.wm.max_level + 1;
 }
 
 static bool intel_wm_plane_visible(const struct intel_crtc_state *crtc_state,
 static void g4x_setup_wm_latency(struct drm_i915_private *dev_priv)
 {
        /* all latencies in usec */
-       dev_priv->wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5;
-       dev_priv->wm.pri_latency[G4X_WM_LEVEL_SR] = 12;
-       dev_priv->wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35;
+       dev_priv->display.wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5;
+       dev_priv->display.wm.pri_latency[G4X_WM_LEVEL_SR] = 12;
+       dev_priv->display.wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35;
 
-       dev_priv->wm.max_level = G4X_WM_LEVEL_HPLL;
+       dev_priv->display.wm.max_level = G4X_WM_LEVEL_HPLL;
 }
 
 static int g4x_plane_fifo_size(enum plane_id plane_id, int level)
        struct drm_i915_private *dev_priv = to_i915(plane->base.dev);
        const struct drm_display_mode *pipe_mode =
                &crtc_state->hw.pipe_mode;
-       unsigned int latency = dev_priv->wm.pri_latency[level] * 10;
+       unsigned int latency = dev_priv->display.wm.pri_latency[level] * 10;
        unsigned int pixel_rate, htotal, cpp, width, wm;
 
        if (latency == 0)
 {
        struct drm_i915_private *dev_priv = to_i915(crtc_state->uapi.crtc->dev);
 
-       if (level > dev_priv->wm.max_level)
+       if (level > dev_priv->display.wm.max_level)
                return false;
 
        return g4x_raw_plane_wm_is_valid(crtc_state, PLANE_PRIMARY, level) &&
 
 static void g4x_program_watermarks(struct drm_i915_private *dev_priv)
 {
-       struct g4x_wm_values *old_wm = &dev_priv->wm.g4x;
+       struct g4x_wm_values *old_wm = &dev_priv->display.wm.g4x;
        struct g4x_wm_values new_wm = {};
 
        g4x_merge_wm(dev_priv, &new_wm);
        const struct intel_crtc_state *crtc_state =
                intel_atomic_get_new_crtc_state(state, crtc);
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
        crtc->wm.active.g4x = crtc_state->wm.g4x.intermediate;
        g4x_program_watermarks(dev_priv);
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 static void g4x_optimize_watermarks(struct intel_atomic_state *state,
        if (!crtc_state->wm.need_postvbl_update)
                return;
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
        crtc->wm.active.g4x = crtc_state->wm.g4x.optimal;
        g4x_program_watermarks(dev_priv);
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 /* latency must be in 0.1us units. */
 static void vlv_setup_wm_latency(struct drm_i915_private *dev_priv)
 {
        /* all latencies in usec */
-       dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM2] = 3;
+       dev_priv->display.wm.pri_latency[VLV_WM_LEVEL_PM2] = 3;
 
-       dev_priv->wm.max_level = VLV_WM_LEVEL_PM2;
+       dev_priv->display.wm.max_level = VLV_WM_LEVEL_PM2;
 
        if (IS_CHERRYVIEW(dev_priv)) {
-               dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM5] = 12;
-               dev_priv->wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33;
+               dev_priv->display.wm.pri_latency[VLV_WM_LEVEL_PM5] = 12;
+               dev_priv->display.wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33;
 
-               dev_priv->wm.max_level = VLV_WM_LEVEL_DDR_DVFS;
+               dev_priv->display.wm.max_level = VLV_WM_LEVEL_DDR_DVFS;
        }
 }
 
                &crtc_state->hw.pipe_mode;
        unsigned int pixel_rate, htotal, cpp, width, wm;
 
-       if (dev_priv->wm.pri_latency[level] == 0)
+       if (dev_priv->display.wm.pri_latency[level] == 0)
                return USHRT_MAX;
 
        if (!intel_wm_plane_visible(crtc_state, plane_state))
                wm = 63;
        } else {
                wm = vlv_wm_method2(pixel_rate, htotal, width, cpp,
-                                   dev_priv->wm.pri_latency[level] * 10);
+                                   dev_priv->display.wm.pri_latency[level] * 10);
        }
 
        return min_t(unsigned int, wm, USHRT_MAX);
        struct intel_crtc *crtc;
        int num_active_pipes = 0;
 
-       wm->level = dev_priv->wm.max_level;
+       wm->level = dev_priv->display.wm.max_level;
        wm->cxsr = true;
 
        for_each_intel_crtc(&dev_priv->drm, crtc) {
 
 static void vlv_program_watermarks(struct drm_i915_private *dev_priv)
 {
-       struct vlv_wm_values *old_wm = &dev_priv->wm.vlv;
+       struct vlv_wm_values *old_wm = &dev_priv->display.wm.vlv;
        struct vlv_wm_values new_wm = {};
 
        vlv_merge_wm(dev_priv, &new_wm);
        const struct intel_crtc_state *crtc_state =
                intel_atomic_get_new_crtc_state(state, crtc);
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
        crtc->wm.active.vlv = crtc_state->wm.vlv.intermediate;
        vlv_program_watermarks(dev_priv);
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 static void vlv_optimize_watermarks(struct intel_atomic_state *state,
        if (!crtc_state->wm.need_postvbl_update)
                return;
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
        crtc->wm.active.vlv = crtc_state->wm.vlv.optimal;
        vlv_program_watermarks(dev_priv);
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 static void i965_update_wm(struct drm_i915_private *dev_priv)
                                 const struct intel_plane_state *curstate,
                                 struct intel_wm_level *result)
 {
-       u16 pri_latency = dev_priv->wm.pri_latency[level];
-       u16 spr_latency = dev_priv->wm.spr_latency[level];
-       u16 cur_latency = dev_priv->wm.cur_latency[level];
+       u16 pri_latency = dev_priv->display.wm.pri_latency[level];
+       u16 spr_latency = dev_priv->display.wm.spr_latency[level];
+       u16 cur_latency = dev_priv->display.wm.cur_latency[level];
 
        /* WM1+ latency values stored in 0.5us units */
        if (level > 0) {
         * The BIOS provided WM memory latency values are often
         * inadequate for high resolution displays. Adjust them.
         */
-       changed = ilk_increase_wm_latency(dev_priv, dev_priv->wm.pri_latency, 12);
-       changed |= ilk_increase_wm_latency(dev_priv, dev_priv->wm.spr_latency, 12);
-       changed |= ilk_increase_wm_latency(dev_priv, dev_priv->wm.cur_latency, 12);
+       changed = ilk_increase_wm_latency(dev_priv, dev_priv->display.wm.pri_latency, 12);
+       changed |= ilk_increase_wm_latency(dev_priv, dev_priv->display.wm.spr_latency, 12);
+       changed |= ilk_increase_wm_latency(dev_priv, dev_priv->display.wm.cur_latency, 12);
 
        if (!changed)
                return;
 
        drm_dbg_kms(&dev_priv->drm,
                    "WM latency values increased to avoid potential underruns\n");
-       intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
-       intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
-       intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
+       intel_print_wm_latency(dev_priv, "Primary", dev_priv->display.wm.pri_latency);
+       intel_print_wm_latency(dev_priv, "Sprite", dev_priv->display.wm.spr_latency);
+       intel_print_wm_latency(dev_priv, "Cursor", dev_priv->display.wm.cur_latency);
 }
 
 static void snb_wm_lp3_irq_quirk(struct drm_i915_private *dev_priv)
         * interrupts only. To play it safe we disable LP3
         * watermarks entirely.
         */
-       if (dev_priv->wm.pri_latency[3] == 0 &&
-           dev_priv->wm.spr_latency[3] == 0 &&
-           dev_priv->wm.cur_latency[3] == 0)
+       if (dev_priv->display.wm.pri_latency[3] == 0 &&
+           dev_priv->display.wm.spr_latency[3] == 0 &&
+           dev_priv->display.wm.cur_latency[3] == 0)
                return;
 
-       dev_priv->wm.pri_latency[3] = 0;
-       dev_priv->wm.spr_latency[3] = 0;
-       dev_priv->wm.cur_latency[3] = 0;
+       dev_priv->display.wm.pri_latency[3] = 0;
+       dev_priv->display.wm.spr_latency[3] = 0;
+       dev_priv->display.wm.cur_latency[3] = 0;
 
        drm_dbg_kms(&dev_priv->drm,
                    "LP3 watermarks disabled due to potential for lost interrupts\n");
-       intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
-       intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
-       intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
+       intel_print_wm_latency(dev_priv, "Primary", dev_priv->display.wm.pri_latency);
+       intel_print_wm_latency(dev_priv, "Sprite", dev_priv->display.wm.spr_latency);
+       intel_print_wm_latency(dev_priv, "Cursor", dev_priv->display.wm.cur_latency);
 }
 
 static void ilk_setup_wm_latency(struct drm_i915_private *dev_priv)
 {
-       intel_read_wm_latency(dev_priv, dev_priv->wm.pri_latency);
+       intel_read_wm_latency(dev_priv, dev_priv->display.wm.pri_latency);
 
-       memcpy(dev_priv->wm.spr_latency, dev_priv->wm.pri_latency,
-              sizeof(dev_priv->wm.pri_latency));
-       memcpy(dev_priv->wm.cur_latency, dev_priv->wm.pri_latency,
-              sizeof(dev_priv->wm.pri_latency));
+       memcpy(dev_priv->display.wm.spr_latency, dev_priv->display.wm.pri_latency,
+              sizeof(dev_priv->display.wm.pri_latency));
+       memcpy(dev_priv->display.wm.cur_latency, dev_priv->display.wm.pri_latency,
+              sizeof(dev_priv->display.wm.pri_latency));
 
-       intel_fixup_spr_wm_latency(dev_priv, dev_priv->wm.spr_latency);
-       intel_fixup_cur_wm_latency(dev_priv, dev_priv->wm.cur_latency);
+       intel_fixup_spr_wm_latency(dev_priv, dev_priv->display.wm.spr_latency);
+       intel_fixup_cur_wm_latency(dev_priv, dev_priv->display.wm.cur_latency);
 
-       intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency);
-       intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency);
-       intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency);
+       intel_print_wm_latency(dev_priv, "Primary", dev_priv->display.wm.pri_latency);
+       intel_print_wm_latency(dev_priv, "Sprite", dev_priv->display.wm.spr_latency);
+       intel_print_wm_latency(dev_priv, "Cursor", dev_priv->display.wm.cur_latency);
 
        if (DISPLAY_VER(dev_priv) == 6) {
                snb_wm_latency_quirk(dev_priv);
 
 static void skl_setup_wm_latency(struct drm_i915_private *dev_priv)
 {
-       intel_read_wm_latency(dev_priv, dev_priv->wm.skl_latency);
-       intel_print_wm_latency(dev_priv, "Gen9 Plane", dev_priv->wm.skl_latency);
+       intel_read_wm_latency(dev_priv, dev_priv->display.wm.skl_latency);
+       intel_print_wm_latency(dev_priv, "Gen9 Plane", dev_priv->display.wm.skl_latency);
 }
 
 static bool ilk_validate_pipe_wm(const struct drm_i915_private *dev_priv,
        if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv))
                return 2 * level;
        else
-               return dev_priv->wm.pri_latency[level];
+               return dev_priv->display.wm.pri_latency[level];
 }
 
 static void ilk_compute_wm_results(struct drm_i915_private *dev_priv,
 static bool _ilk_disable_lp_wm(struct drm_i915_private *dev_priv,
                               unsigned int dirty)
 {
-       struct ilk_wm_values *previous = &dev_priv->wm.hw;
+       struct ilk_wm_values *previous = &dev_priv->display.wm.hw;
        bool changed = false;
 
        if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] & WM_LP_ENABLE) {
 static void ilk_write_wm_values(struct drm_i915_private *dev_priv,
                                struct ilk_wm_values *results)
 {
-       struct ilk_wm_values *previous = &dev_priv->wm.hw;
+       struct ilk_wm_values *previous = &dev_priv->display.wm.hw;
        unsigned int dirty;
        u32 val;
 
        if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] != results->wm_lp[2])
                intel_uncore_write(&dev_priv->uncore, WM3_LP_ILK, results->wm_lp[2]);
 
-       dev_priv->wm.hw = *results;
+       dev_priv->display.wm.hw = *results;
 }
 
 bool ilk_disable_lp_wm(struct drm_i915_private *dev_priv)
        drm_WARN_ON(&dev_priv->drm, ret);
 
        for (level = 0; level <= max_level; level++) {
-               unsigned int latency = dev_priv->wm.skl_latency[level];
+               unsigned int latency = dev_priv->display.wm.skl_latency[level];
 
                skl_compute_plane_wm(crtc_state, plane, level, latency, &wp, &wm, &wm);
                if (wm.min_ddb_alloc == U16_MAX)
 
        for (level = 0; level <= max_level; level++) {
                struct skl_wm_level *result = &levels[level];
-               unsigned int latency = dev_priv->wm.skl_latency[level];
+               unsigned int latency = dev_priv->display.wm.skl_latency[level];
 
                skl_compute_plane_wm(crtc_state, plane, level, latency,
                                     wm_params, result_prev, result);
        unsigned int latency = 0;
 
        if (dev_priv->sagv_block_time_us)
-               latency = dev_priv->sagv_block_time_us + dev_priv->wm.skl_latency[0];
+               latency = dev_priv->sagv_block_time_us + dev_priv->display.wm.skl_latency[0];
 
        skl_compute_plane_wm(crtc_state, plane, 0, latency,
                             wm_params, &levels[0],
        const struct intel_crtc_state *crtc_state =
                intel_atomic_get_new_crtc_state(state, crtc);
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
        crtc->wm.active.ilk = crtc_state->wm.ilk.intermediate;
        ilk_program_watermarks(dev_priv);
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 static void ilk_optimize_watermarks(struct intel_atomic_state *state,
        if (!crtc_state->wm.need_postvbl_update)
                return;
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
        crtc->wm.active.ilk = crtc_state->wm.ilk.optimal;
        ilk_program_watermarks(dev_priv);
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 static void skl_wm_level_from_reg_val(u32 val, struct skl_wm_level *level)
 {
        struct drm_device *dev = crtc->base.dev;
        struct drm_i915_private *dev_priv = to_i915(dev);
-       struct ilk_wm_values *hw = &dev_priv->wm.hw;
+       struct ilk_wm_values *hw = &dev_priv->display.wm.hw;
        struct intel_crtc_state *crtc_state = to_intel_crtc_state(crtc->base.state);
        struct intel_pipe_wm *active = &crtc_state->wm.ilk.optimal;
        enum pipe pipe = crtc->pipe;
 
 void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv)
 {
-       struct g4x_wm_values *wm = &dev_priv->wm.g4x;
+       struct g4x_wm_values *wm = &dev_priv->display.wm.g4x;
        struct intel_crtc *crtc;
 
        g4x_read_wm_values(dev_priv, wm);
        struct intel_plane *plane;
        struct intel_crtc *crtc;
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
 
        for_each_intel_plane(&dev_priv->drm, plane) {
                struct intel_crtc *crtc =
 
        g4x_program_watermarks(dev_priv);
 
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv)
 {
-       struct vlv_wm_values *wm = &dev_priv->wm.vlv;
+       struct vlv_wm_values *wm = &dev_priv->display.wm.vlv;
        struct intel_crtc *crtc;
        u32 val;
 
                        drm_dbg_kms(&dev_priv->drm,
                                    "Punit not acking DDR DVFS request, "
                                    "assuming DDR DVFS is disabled\n");
-                       dev_priv->wm.max_level = VLV_WM_LEVEL_PM5;
+                       dev_priv->display.wm.max_level = VLV_WM_LEVEL_PM5;
                } else {
                        val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2);
                        if ((val & FORCE_DDR_HIGH_FREQ) == 0)
        struct intel_plane *plane;
        struct intel_crtc *crtc;
 
-       mutex_lock(&dev_priv->wm.wm_mutex);
+       mutex_lock(&dev_priv->display.wm.wm_mutex);
 
        for_each_intel_plane(&dev_priv->drm, plane) {
                struct intel_crtc *crtc =
 
        vlv_program_watermarks(dev_priv);
 
-       mutex_unlock(&dev_priv->wm.wm_mutex);
+       mutex_unlock(&dev_priv->display.wm.wm_mutex);
 }
 
 /*
 
 void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv)
 {
-       struct ilk_wm_values *hw = &dev_priv->wm.hw;
+       struct ilk_wm_values *hw = &dev_priv->display.wm.hw;
        struct intel_crtc *crtc;
 
        ilk_init_lp_watermarks(dev_priv);
        } else if (HAS_PCH_SPLIT(dev_priv)) {
                ilk_setup_wm_latency(dev_priv);
 
-               if ((DISPLAY_VER(dev_priv) == 5 && dev_priv->wm.pri_latency[1] &&
-                    dev_priv->wm.spr_latency[1] && dev_priv->wm.cur_latency[1]) ||
-                   (DISPLAY_VER(dev_priv) != 5 && dev_priv->wm.pri_latency[0] &&
-                    dev_priv->wm.spr_latency[0] && dev_priv->wm.cur_latency[0])) {
+               if ((DISPLAY_VER(dev_priv) == 5 && dev_priv->display.wm.pri_latency[1] &&
+                    dev_priv->display.wm.spr_latency[1] && dev_priv->display.wm.cur_latency[1]) ||
+                   (DISPLAY_VER(dev_priv) != 5 && dev_priv->display.wm.pri_latency[0] &&
+                    dev_priv->display.wm.spr_latency[0] && dev_priv->display.wm.cur_latency[0])) {
                        dev_priv->display.funcs.wm = &ilk_wm_funcs;
                } else {
                        drm_dbg_kms(&dev_priv->drm,