drm/msm/dsi: make save_state/restore_state callbacks accept msm_dsi_phy
Make save_state/restore callbacks accept struct msm_dsi_phy rather than struct msm_dsi_pll. This moves them to struct msm_dsi_phy_ops, allowing us to drop struct msm_dsi_pll_ops. Signed-off-by: Dmitry Baryshkov <dmitry.baryshkov@linaro.org> Tested-by: Stephen Boyd <swboyd@chromium.org> # on sc7180 lazor Reviewed-by: Abhinav Kumar <abhinavk@codeaurora.org> Link: https://lore.kernel.org/r/20210331105735.3690009-18-dmitry.baryshkov@linaro.org Signed-off-by: Rob Clark <robdclark@chromium.org>
This commit is contained in:
committed by
Rob Clark
parent
62d5325d45
commit
2a831d9e38
@@ -858,9 +858,9 @@ int msm_dsi_phy_get_clk_provider(struct msm_dsi_phy *phy,
|
||||
|
||||
void msm_dsi_phy_pll_save_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
if (phy->cfg->pll_ops.save_state) {
|
||||
phy->cfg->pll_ops.save_state(phy->pll);
|
||||
phy->pll->state_saved = true;
|
||||
if (phy->cfg->ops.save_pll_state) {
|
||||
phy->cfg->ops.save_pll_state(phy);
|
||||
phy->state_saved = true;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -868,12 +868,12 @@ int msm_dsi_phy_pll_restore_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
int ret;
|
||||
|
||||
if (phy->cfg->pll_ops.restore_state && phy->pll->state_saved) {
|
||||
ret = phy->cfg->pll_ops.restore_state(phy->pll);
|
||||
if (phy->cfg->ops.restore_pll_state && phy->state_saved) {
|
||||
ret = phy->cfg->ops.restore_pll_state(phy);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
phy->pll->state_saved = false;
|
||||
phy->state_saved = false;
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
||||
@@ -17,7 +17,6 @@
|
||||
struct msm_dsi_pll {
|
||||
struct clk_hw clk_hw;
|
||||
bool pll_on;
|
||||
bool state_saved;
|
||||
|
||||
const struct msm_dsi_phy_cfg *cfg;
|
||||
};
|
||||
@@ -29,17 +28,13 @@ struct msm_dsi_phy_ops {
|
||||
int (*enable)(struct msm_dsi_phy *phy, int src_pll_id,
|
||||
struct msm_dsi_phy_clk_request *clk_req);
|
||||
void (*disable)(struct msm_dsi_phy *phy);
|
||||
};
|
||||
|
||||
struct msm_dsi_pll_ops {
|
||||
void (*save_state)(struct msm_dsi_pll *pll);
|
||||
int (*restore_state)(struct msm_dsi_pll *pll);
|
||||
void (*save_pll_state)(struct msm_dsi_phy *phy);
|
||||
int (*restore_pll_state)(struct msm_dsi_phy *phy);
|
||||
};
|
||||
|
||||
struct msm_dsi_phy_cfg {
|
||||
struct dsi_reg_config reg_cfg;
|
||||
struct msm_dsi_phy_ops ops;
|
||||
const struct msm_dsi_pll_ops pll_ops;
|
||||
|
||||
unsigned long min_pll_rate;
|
||||
unsigned long max_pll_rate;
|
||||
@@ -115,6 +110,8 @@ struct msm_dsi_phy {
|
||||
struct msm_dsi_pll *pll;
|
||||
|
||||
struct clk_hw_onecell_data *provided_clocks;
|
||||
|
||||
bool state_saved;
|
||||
};
|
||||
|
||||
/*
|
||||
|
||||
@@ -518,9 +518,9 @@ static const struct clk_ops clk_ops_dsi_pll_10nm_vco = {
|
||||
* PLL Callbacks
|
||||
*/
|
||||
|
||||
static void dsi_pll_10nm_save_state(struct msm_dsi_pll *pll)
|
||||
static void dsi_10nm_pll_save_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_10nm *pll_10nm = to_pll_10nm(pll);
|
||||
struct dsi_pll_10nm *pll_10nm = to_pll_10nm(phy->pll);
|
||||
struct pll_10nm_cached_state *cached = &pll_10nm->cached_state;
|
||||
void __iomem *phy_base = pll_10nm->phy_cmn_mmio;
|
||||
u32 cmn_clk_cfg0, cmn_clk_cfg1;
|
||||
@@ -541,9 +541,9 @@ static void dsi_pll_10nm_save_state(struct msm_dsi_pll *pll)
|
||||
cached->pix_clk_div, cached->pll_mux);
|
||||
}
|
||||
|
||||
static int dsi_pll_10nm_restore_state(struct msm_dsi_pll *pll)
|
||||
static int dsi_10nm_pll_restore_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_10nm *pll_10nm = to_pll_10nm(pll);
|
||||
struct dsi_pll_10nm *pll_10nm = to_pll_10nm(phy->pll);
|
||||
struct pll_10nm_cached_state *cached = &pll_10nm->cached_state;
|
||||
void __iomem *phy_base = pll_10nm->phy_cmn_mmio;
|
||||
u32 val;
|
||||
@@ -562,7 +562,9 @@ static int dsi_pll_10nm_restore_state(struct msm_dsi_pll *pll)
|
||||
val |= cached->pll_mux;
|
||||
pll_write(phy_base + REG_DSI_10nm_PHY_CMN_CLK_CFG1, val);
|
||||
|
||||
ret = dsi_pll_10nm_vco_set_rate(&pll->clk_hw, pll_10nm->vco_current_rate, pll_10nm->vco_ref_clk_rate);
|
||||
ret = dsi_pll_10nm_vco_set_rate(&phy->pll->clk_hw,
|
||||
pll_10nm->vco_current_rate,
|
||||
pll_10nm->vco_ref_clk_rate);
|
||||
if (ret) {
|
||||
DRM_DEV_ERROR(&pll_10nm->pdev->dev,
|
||||
"restore vco rate failed. ret=%d\n", ret);
|
||||
@@ -1005,10 +1007,8 @@ const struct msm_dsi_phy_cfg dsi_phy_10nm_cfgs = {
|
||||
.enable = dsi_10nm_phy_enable,
|
||||
.disable = dsi_10nm_phy_disable,
|
||||
.pll_init = dsi_pll_10nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_10nm_save_state,
|
||||
.restore_state = dsi_pll_10nm_restore_state,
|
||||
.save_pll_state = dsi_10nm_pll_save_state,
|
||||
.restore_pll_state = dsi_10nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = 1000000000UL,
|
||||
.max_pll_rate = 3500000000UL,
|
||||
@@ -1029,10 +1029,8 @@ const struct msm_dsi_phy_cfg dsi_phy_10nm_8998_cfgs = {
|
||||
.enable = dsi_10nm_phy_enable,
|
||||
.disable = dsi_10nm_phy_disable,
|
||||
.pll_init = dsi_pll_10nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_10nm_save_state,
|
||||
.restore_state = dsi_pll_10nm_restore_state,
|
||||
.save_pll_state = dsi_10nm_pll_save_state,
|
||||
.restore_pll_state = dsi_10nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = 1000000000UL,
|
||||
.max_pll_rate = 3500000000UL,
|
||||
|
||||
@@ -795,9 +795,9 @@ static const struct clk_ops clk_ops_dsi_pll_14nm_postdiv = {
|
||||
* PLL Callbacks
|
||||
*/
|
||||
|
||||
static void dsi_pll_14nm_save_state(struct msm_dsi_pll *pll)
|
||||
static void dsi_14nm_pll_save_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_14nm *pll_14nm = to_pll_14nm(pll);
|
||||
struct dsi_pll_14nm *pll_14nm = to_pll_14nm(phy->pll);
|
||||
struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state;
|
||||
void __iomem *cmn_base = pll_14nm->phy_cmn_mmio;
|
||||
u32 data;
|
||||
@@ -810,18 +810,18 @@ static void dsi_pll_14nm_save_state(struct msm_dsi_pll *pll)
|
||||
DBG("DSI%d PLL save state %x %x", pll_14nm->id,
|
||||
cached_state->n1postdiv, cached_state->n2postdiv);
|
||||
|
||||
cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw);
|
||||
cached_state->vco_rate = clk_hw_get_rate(&phy->pll->clk_hw);
|
||||
}
|
||||
|
||||
static int dsi_pll_14nm_restore_state(struct msm_dsi_pll *pll)
|
||||
static int dsi_14nm_pll_restore_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_14nm *pll_14nm = to_pll_14nm(pll);
|
||||
struct dsi_pll_14nm *pll_14nm = to_pll_14nm(phy->pll);
|
||||
struct pll_14nm_cached_state *cached_state = &pll_14nm->cached_state;
|
||||
void __iomem *cmn_base = pll_14nm->phy_cmn_mmio;
|
||||
u32 data;
|
||||
int ret;
|
||||
|
||||
ret = dsi_pll_14nm_vco_set_rate(&pll->clk_hw,
|
||||
ret = dsi_pll_14nm_vco_set_rate(&phy->pll->clk_hw,
|
||||
cached_state->vco_rate, 0);
|
||||
if (ret) {
|
||||
DRM_DEV_ERROR(&pll_14nm->pdev->dev,
|
||||
@@ -1166,10 +1166,8 @@ const struct msm_dsi_phy_cfg dsi_phy_14nm_cfgs = {
|
||||
.enable = dsi_14nm_phy_enable,
|
||||
.disable = dsi_14nm_phy_disable,
|
||||
.pll_init = dsi_pll_14nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_14nm_save_state,
|
||||
.restore_state = dsi_pll_14nm_restore_state,
|
||||
.save_pll_state = dsi_14nm_pll_save_state,
|
||||
.restore_pll_state = dsi_14nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = VCO_MIN_RATE,
|
||||
.max_pll_rate = VCO_MAX_RATE,
|
||||
@@ -1190,10 +1188,8 @@ const struct msm_dsi_phy_cfg dsi_phy_14nm_660_cfgs = {
|
||||
.enable = dsi_14nm_phy_enable,
|
||||
.disable = dsi_14nm_phy_disable,
|
||||
.pll_init = dsi_pll_14nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_14nm_save_state,
|
||||
.restore_state = dsi_pll_14nm_restore_state,
|
||||
.save_pll_state = dsi_14nm_pll_save_state,
|
||||
.restore_pll_state = dsi_14nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = VCO_MIN_RATE,
|
||||
.max_pll_rate = VCO_MAX_RATE,
|
||||
|
||||
@@ -470,9 +470,9 @@ static const struct clk_ops clk_ops_dsi_pll_28nm_vco_lp = {
|
||||
* PLL Callbacks
|
||||
*/
|
||||
|
||||
static void dsi_pll_28nm_save_state(struct msm_dsi_pll *pll)
|
||||
static void dsi_28nm_pll_save_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(pll);
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(phy->pll);
|
||||
struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state;
|
||||
void __iomem *base = pll_28nm->mmio;
|
||||
|
||||
@@ -481,20 +481,20 @@ static void dsi_pll_28nm_save_state(struct msm_dsi_pll *pll)
|
||||
cached_state->postdiv1 =
|
||||
pll_read(base + REG_DSI_28nm_PHY_PLL_POSTDIV1_CFG);
|
||||
cached_state->byte_mux = pll_read(base + REG_DSI_28nm_PHY_PLL_VREG_CFG);
|
||||
if (dsi_pll_28nm_clk_is_enabled(&pll->clk_hw))
|
||||
cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw);
|
||||
if (dsi_pll_28nm_clk_is_enabled(&phy->pll->clk_hw))
|
||||
cached_state->vco_rate = clk_hw_get_rate(&phy->pll->clk_hw);
|
||||
else
|
||||
cached_state->vco_rate = 0;
|
||||
}
|
||||
|
||||
static int dsi_pll_28nm_restore_state(struct msm_dsi_pll *pll)
|
||||
static int dsi_28nm_pll_restore_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(pll);
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(phy->pll);
|
||||
struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state;
|
||||
void __iomem *base = pll_28nm->mmio;
|
||||
int ret;
|
||||
|
||||
ret = dsi_pll_28nm_clk_set_rate(&pll->clk_hw,
|
||||
ret = dsi_pll_28nm_clk_set_rate(&phy->pll->clk_hw,
|
||||
cached_state->vco_rate, 0);
|
||||
if (ret) {
|
||||
DRM_DEV_ERROR(&pll_28nm->pdev->dev,
|
||||
@@ -527,7 +527,7 @@ static int pll_28nm_register(struct dsi_pll_28nm *pll_28nm, struct clk_hw **prov
|
||||
|
||||
DBG("%d", pll_28nm->id);
|
||||
|
||||
if (pll_28nm->base.cfg->type == MSM_DSI_PHY_28NM_LP)
|
||||
if (pll_28nm->base.cfg->quirks & DSI_PHY_28NM_QUIRK_PHY_LP)
|
||||
vco_init.ops = &clk_ops_dsi_pll_28nm_vco_lp;
|
||||
else
|
||||
vco_init.ops = &clk_ops_dsi_pll_28nm_vco_hpm;
|
||||
@@ -783,10 +783,8 @@ const struct msm_dsi_phy_cfg dsi_phy_28nm_hpm_cfgs = {
|
||||
.enable = dsi_28nm_phy_enable,
|
||||
.disable = dsi_28nm_phy_disable,
|
||||
.pll_init = dsi_pll_28nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_28nm_save_state,
|
||||
.restore_state = dsi_pll_28nm_restore_state,
|
||||
.save_pll_state = dsi_28nm_pll_save_state,
|
||||
.restore_pll_state = dsi_28nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = VCO_MIN_RATE,
|
||||
.max_pll_rate = VCO_MAX_RATE,
|
||||
@@ -807,10 +805,8 @@ const struct msm_dsi_phy_cfg dsi_phy_28nm_hpm_famb_cfgs = {
|
||||
.enable = dsi_28nm_phy_enable,
|
||||
.disable = dsi_28nm_phy_disable,
|
||||
.pll_init = dsi_pll_28nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_28nm_save_state,
|
||||
.restore_state = dsi_pll_28nm_restore_state,
|
||||
.save_pll_state = dsi_28nm_pll_save_state,
|
||||
.restore_pll_state = dsi_28nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = VCO_MIN_RATE,
|
||||
.max_pll_rate = VCO_MAX_RATE,
|
||||
@@ -831,10 +827,8 @@ const struct msm_dsi_phy_cfg dsi_phy_28nm_lp_cfgs = {
|
||||
.enable = dsi_28nm_phy_enable,
|
||||
.disable = dsi_28nm_phy_disable,
|
||||
.pll_init = dsi_pll_28nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_28nm_save_state,
|
||||
.restore_state = dsi_pll_28nm_restore_state,
|
||||
.save_pll_state = dsi_28nm_pll_save_state,
|
||||
.restore_pll_state = dsi_28nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = VCO_MIN_RATE,
|
||||
.max_pll_rate = VCO_MAX_RATE,
|
||||
|
||||
@@ -334,9 +334,9 @@ static const struct clk_ops clk_bytediv_ops = {
|
||||
/*
|
||||
* PLL Callbacks
|
||||
*/
|
||||
static void dsi_pll_28nm_save_state(struct msm_dsi_pll *pll)
|
||||
static void dsi_28nm_pll_save_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(pll);
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(phy->pll);
|
||||
struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state;
|
||||
void __iomem *base = pll_28nm->mmio;
|
||||
|
||||
@@ -347,17 +347,17 @@ static void dsi_pll_28nm_save_state(struct msm_dsi_pll *pll)
|
||||
cached_state->postdiv1 =
|
||||
pll_read(base + REG_DSI_28nm_8960_PHY_PLL_CTRL_8);
|
||||
|
||||
cached_state->vco_rate = clk_hw_get_rate(&pll->clk_hw);
|
||||
cached_state->vco_rate = clk_hw_get_rate(&phy->pll->clk_hw);
|
||||
}
|
||||
|
||||
static int dsi_pll_28nm_restore_state(struct msm_dsi_pll *pll)
|
||||
static int dsi_28nm_pll_restore_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(pll);
|
||||
struct dsi_pll_28nm *pll_28nm = to_pll_28nm(phy->pll);
|
||||
struct pll_28nm_cached_state *cached_state = &pll_28nm->cached_state;
|
||||
void __iomem *base = pll_28nm->mmio;
|
||||
int ret;
|
||||
|
||||
ret = dsi_pll_28nm_clk_set_rate(&pll->clk_hw,
|
||||
ret = dsi_pll_28nm_clk_set_rate(&phy->pll->clk_hw,
|
||||
cached_state->vco_rate, 0);
|
||||
if (ret) {
|
||||
DRM_DEV_ERROR(&pll_28nm->pdev->dev,
|
||||
@@ -662,10 +662,8 @@ const struct msm_dsi_phy_cfg dsi_phy_28nm_8960_cfgs = {
|
||||
.enable = dsi_28nm_phy_enable,
|
||||
.disable = dsi_28nm_phy_disable,
|
||||
.pll_init = dsi_pll_28nm_8960_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_28nm_save_state,
|
||||
.restore_state = dsi_pll_28nm_restore_state,
|
||||
.save_pll_state = dsi_28nm_pll_save_state,
|
||||
.restore_pll_state = dsi_28nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = VCO_MIN_RATE,
|
||||
.max_pll_rate = VCO_MAX_RATE,
|
||||
|
||||
@@ -543,9 +543,9 @@ static const struct clk_ops clk_ops_dsi_pll_7nm_vco = {
|
||||
* PLL Callbacks
|
||||
*/
|
||||
|
||||
static void dsi_pll_7nm_save_state(struct msm_dsi_pll *pll)
|
||||
static void dsi_7nm_pll_save_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_7nm *pll_7nm = to_pll_7nm(pll);
|
||||
struct dsi_pll_7nm *pll_7nm = to_pll_7nm(phy->pll);
|
||||
struct pll_7nm_cached_state *cached = &pll_7nm->cached_state;
|
||||
void __iomem *phy_base = pll_7nm->phy_cmn_mmio;
|
||||
u32 cmn_clk_cfg0, cmn_clk_cfg1;
|
||||
@@ -566,9 +566,9 @@ static void dsi_pll_7nm_save_state(struct msm_dsi_pll *pll)
|
||||
cached->pix_clk_div, cached->pll_mux);
|
||||
}
|
||||
|
||||
static int dsi_pll_7nm_restore_state(struct msm_dsi_pll *pll)
|
||||
static int dsi_7nm_pll_restore_state(struct msm_dsi_phy *phy)
|
||||
{
|
||||
struct dsi_pll_7nm *pll_7nm = to_pll_7nm(pll);
|
||||
struct dsi_pll_7nm *pll_7nm = to_pll_7nm(phy->pll);
|
||||
struct pll_7nm_cached_state *cached = &pll_7nm->cached_state;
|
||||
void __iomem *phy_base = pll_7nm->phy_cmn_mmio;
|
||||
u32 val;
|
||||
@@ -587,7 +587,9 @@ static int dsi_pll_7nm_restore_state(struct msm_dsi_pll *pll)
|
||||
val |= cached->pll_mux;
|
||||
pll_write(phy_base + REG_DSI_7nm_PHY_CMN_CLK_CFG1, val);
|
||||
|
||||
ret = dsi_pll_7nm_vco_set_rate(&pll->clk_hw, pll_7nm->vco_current_rate, pll_7nm->vco_ref_clk_rate);
|
||||
ret = dsi_pll_7nm_vco_set_rate(&phy->pll->clk_hw,
|
||||
pll_7nm->vco_current_rate,
|
||||
pll_7nm->vco_ref_clk_rate);
|
||||
if (ret) {
|
||||
DRM_DEV_ERROR(&pll_7nm->pdev->dev,
|
||||
"restore vco rate failed. ret=%d\n", ret);
|
||||
@@ -1038,10 +1040,8 @@ const struct msm_dsi_phy_cfg dsi_phy_7nm_cfgs = {
|
||||
.enable = dsi_7nm_phy_enable,
|
||||
.disable = dsi_7nm_phy_disable,
|
||||
.pll_init = dsi_pll_7nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_7nm_save_state,
|
||||
.restore_state = dsi_pll_7nm_restore_state,
|
||||
.save_pll_state = dsi_7nm_pll_save_state,
|
||||
.restore_pll_state = dsi_7nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = 600000000UL,
|
||||
.max_pll_rate = (5000000000ULL < ULONG_MAX) ? 5000000000ULL : ULONG_MAX,
|
||||
@@ -1063,10 +1063,8 @@ const struct msm_dsi_phy_cfg dsi_phy_7nm_8150_cfgs = {
|
||||
.enable = dsi_7nm_phy_enable,
|
||||
.disable = dsi_7nm_phy_disable,
|
||||
.pll_init = dsi_pll_7nm_init,
|
||||
},
|
||||
.pll_ops = {
|
||||
.save_state = dsi_pll_7nm_save_state,
|
||||
.restore_state = dsi_pll_7nm_restore_state,
|
||||
.save_pll_state = dsi_7nm_pll_save_state,
|
||||
.restore_pll_state = dsi_7nm_pll_restore_state,
|
||||
},
|
||||
.min_pll_rate = 1000000000UL,
|
||||
.max_pll_rate = 3500000000UL,
|
||||
|
||||
Reference in New Issue
Block a user