楼主: 小楼明月
13915 7

[问答] [求助]是否应该考虑控制变量的多重共线性问题? [推广有奖]

  • 0关注
  • 0粉丝

学前班

80%

还不是VIP/贵宾

-

威望
0
论坛币
10 个
通用积分
0
学术水平
0 点
热心指数
0 点
信用等级
0 点
经验
70 点
帖子
2
精华
0
在线时间
0 小时
注册时间
2009-2-26
最后登录
2009-2-26

楼主
小楼明月 发表于 2009-2-26 09:04:00 |AI写论文

+2 论坛币
k人 参与回答

经管之家送您一份

应届毕业生专属福利!

求职就业群
赵安豆老师微信:zhaoandou666

经管之家联合CDA

送您一个全额奖学金名额~ !

感谢您参与论坛问题回答

经管之家送您两个论坛币!

+2 论坛币

论坛上有过关于控制变量的问题,有人说重来没听说过控制变量.........我是写财务会计方向的论文的,经常触到这个概念,我女朋友学人力资源,他们做研究也经常用到这个....可是我对这个概念理解不是很透彻,只知道是控制其它因素对被解释变量的影响.....我现在的问题是:需不需要考虑控制变量的多重共线性呢??

比如我的拟定的模型中有控制变量资产报酬率(ROA),也有总资产,还有资产负债率,无形资产比,这不很明显会共线性嘛.....

财务指标之间的关系太过密切,如果要考虑控制变量的共线性,真不知道应该怎么办了......各位高手救命啊.....

另外:SPSS中怎么添加控制变量呢?

二维码

扫码加我 拉你入群

请注明:姓名-公司-职位

以便审核进群资格,未注明则拒绝

关键词:多重共线性问题 多重共线性 是否应该 控制变量 多重共线 线性 控制变量

沙发
凌子墨 发表于 2009-2-26 09:26:00

多重共线性会影响估计结果的精确度,即方差变大。解决的办法是增加样本量,不过一般很难实现。所以,共线性能降低最好,实在不行也不必强求。

但有一点要注意,完全共线性是不可以的,违反了满秩条件,无法估计。

具体可参照伍德里奇的《现代观点》,里面说的很清楚

世上最遥远的距离,不是生与死的距离,不是天各一方,而是我就站在你面前,你却不知道我爱你。——张小娴

藤椅
欣华123 发表于 2013-7-29 20:43:31
请问楼主这个问题现在解决了没,我现在也很想知道,如果有幸被楼主看到,请告知,请高手救命~

板凳
流潋 在职认证  发表于 2013-8-6 19:16:47
同楼上,这个问题要如何解决呢?

报纸
lovewcq123 发表于 2016-4-3 09:38:32
http://statisticalhorizons.com/multicollinearity

Regardless of your criterion for what constitutes a high VIF, there are at least three situations in which a high VIF is not a problem and can be safely ignored:

1. The variables with high VIFs are control variables, and the variables of interest do not have high VIFs. Here’s the thing about multicollinearity: it’s only a problem for the variables that are collinear. It increases the standard errors of their coefficients, and it may make those coefficients unstable in several ways. But so long as the collinear variables are only used as control variables, and they are not collinear with your variables of interest, there’s no problem. The coefficients of the variables of interest are not affected, and the performance of the control variables as controls is not impaired.

Here’s an example from some of my own work: the sample consists of U.S. colleges, the dependent variable is graduation rate, and the variable of interest is an indicator (dummy) for public vs. private. Two control variables are average SAT scores and average ACT scores for entering freshmen. These two variables have a correlation above .9, which corresponds to VIFs of at least 5.26 for each of them. But the VIF for the public/private indicator is only 1.04. So there’s no problem to be concerned about, and no need to delete one or the other of the two controls.

2. The high VIFs are caused by the inclusion of powers or products of other variables. If you specify a regression model with both x and x2, there’s a good chance that those two variables will be highly correlated. Similarly, if your model has x, z, and xz, both x and z are likely to be highly correlated with their product. This is not something to be concerned about, however, because the p-value for xz is not affected by the multicollinearity.  This is easily demonstrated: you can greatly reduce the correlations by “centering” the variables (i.e., subtracting their means) before creating the powers or the products. But the p-value for x2 or for xz will be exactly the same, regardless of whether or not you center. And all the results for the other variables (including the R2 but not including the lower-order terms) will be the same in either case. So the multicollinearity has no adverse consequences.

3. The variables with high VIFs are indicator (dummy) variables that represent a categorical variable with three or more categories. If the proportion of cases in the reference category is small, the indicator variables will necessarily have high VIFs, even if the categorical variable is not associated with other variables in the regression model.

Suppose, for example, that a marital status variable has three categories: currently married, never married, and formerly married. You choose formerly married as the reference category, with indicator variables for the other two. What happens is that the correlation between those two indicators gets more negative as the fraction of people in the reference category gets smaller. For example, if 45 percent of people are never married, 45 percent are married, and 10 percent are formerly married, the VIFs for the married and never-married indicators will be at least 3.0.

Is this a problem? Well, it does mean that p-values for the indicator variables may be high. But the overall test that all indicators have coefficients of zero is unaffected by the high VIFs. And nothing else in the regression is affected. If you really want to avoid the high VIFs, just choose a reference category with a larger fraction of the cases. That may be desirable in order to avoid situations where none of the individual indicators is statistically significant even though the overall set of indicators is significant.

地板
祥夫孙 发表于 2020-6-15 11:18:09 来自手机
提示: 该帖被管理员或版主屏蔽  gloryfly 灌水 2021-12-30 01:01

7
祥夫孙 发表于 2020-6-15 11:19:41 来自手机
提示: 该帖被管理员或版主屏蔽  gloryfly 灌水 2021-12-30 01:01

8
zydyyds 发表于 2023-4-6 12:00:57
2023年了,我仍有这个疑问。。。

您需要登录后才可以回帖 登录 | 我要注册

本版微信群
加好友,备注cda
拉您进交流群
GMT+8, 2025-12-9 07:01