ASA Connect

 View Only
  • 1.  Advice on the validity of an analysis, please.

    Posted 07-11-2016 14:10

    Dear All,

      

    I’m having an internal argument between myself and me over whether an analysis is or is not valid and I’d appreciate any thoughts, feedback, and suggestions that anyone would be willing to offer.  

      

    I’m very concerned about the validity of the analysis described below because of low sample size and because the dependent variable is the sum of the 5 different independent variables. 

    The study participants are grouped into three groups.  The sample size for the pooled participants is greater than 30 but not by much.  Consequently, the sample sizes in the groups are low.  Other than coding of Likert responses, I haven’t done any weighting or other manipulations and transformations of the data.

      

    The participants complete a 24 item Likert scale questionnaire.  Each Likert item has 5 possible answers, coded 1 through 5.  The 24 items are grouped into 5 subscales and, for each participant, a subscale score is calculated by summing the coded answers for the questions in each subscale (5 subscale scores per participant).  Subscale 1 has 7 items (range of possible subscale scores, 7 to 35), subscale 2 has 6 items (range of possible subscale scores, 6 to 30), subscale 3 has 5 items (range of possible subscale scores, 5 to 25), and subscale 4 and 5 each have 3 items (range of possible subscale scores, 3 to 15).  Finally, an overall score is calculated for each participant by summing the answers on all 24 questions (or by summing the 5 subscale scores).

      

    I have a hypothesis (more of a hunch, really) that, for different groups, scores on different subscales are differentially “contributing” to increasing the overall score for the questionnaire as a whole.  In other words, in group 1, subscale score 4 may be the main “contributor” to the overall Likert questionnaire sum.  However, in group 2, subscale scores 1 and 4 may be the main “contributors” to the overall Likert questionnaire sum.  In group 3, subscale scores 1, 2, and 4 may be the main “contributors” to the overall Likert questionnaire sum.  

      

    To test my hypothesis, for each group, I ran five separate bivariate Pearson correlation analyses (a total of 15 analyses all told) for each combination of the one of the scores on the five subscales (5 independent variables) and the single overall score (dependent variable).  The correlation coefficients support my hypothesis/hunch that different scores on different subscales are differentially contributing to the overall score for the questionnaire as a whole.  Depending on group of participants, not all subscale scores are correlated statistically significantly with the overall score.  In other cases, depending on group, some subscale scores are marginally statistically significantly correlated (p values between .01 and .049).  I’m not convinced that these are real differences so I’m tending to discount them.  But, in some cases, the correlations are highly significant (rho > .9, p values < .005).  These might be real differences.  Also, the patterns of high significance differ between groups (i.e., a given subscale score may be highly correlated with the overall score for one group but not for another group).  This tends to support my hypothesis/hunch.

      

    However, I’m very concerned (approaching convinced) that this analysis is not valid because (1) the sample sizes within the participant groups is so low and (2) the dependent variable (the overall score for each participant) is a sum of the five independent variables (the scores on the five subscales for each participant).  On the good side: The answers of participants are independent observations.  The post hoc effect sizes and the post hoc power of the statistically significant results are high.  The measurement scales of the variables are appropriate for the test.  But this nagging doubt continues…

      

    In order to quiet the inner disagreement between myself and me, I’d appreciate any thoughts, feedback, and suggestions that anyone would be willing to offer, particularly for alternative and more appropriate analysis strategies if any might exist (maybe none exist....).

      

    Thanks in advance for your time and responses.

      

    Linda

    ------------------------------
    Linda A. Landon, PhD, ELS
    Research Communiqué

    Clear, Concise Statistics & Words

    PhD, Molecular Pharmacology
    Graduate Certificate, Applied Statistics
    Board-Certified Editor in the Life Sciences

    Email: LandonPhD@ResearchCommunique.com
    Phone: 573-797-4517
    Website: www.researchcommunique.com
    Jefferson City, MO
    Central Standard Time ( CST ) = GMT-6 (November – February)
    Central Daylight Time ( CDT ) = GMT-5 (March –October)
    ------------------------------


  • 2.  RE: Advice on the validity of an analysis, please.

    Posted 07-12-2016 12:27

    Some thoughts about your data set and question:

    1. All else equal, differences in group or individual status on a summated score are proportional to the standard deviations of the elements summed.  So one obvious point is, the scales having more items are more likely to have 'stronger influence' on the total score than are scales with fewer items. I'd suggest replacing scale scores as sums with scale scores as means of the responses.  That is, divide your scale scores by the number of items contributing to them (e.g., average_1 = scale_1 / 7).  That at least adjusts for the artificial effect of number of items.

    2. Given #1, the next question is, are groups comparable in variance on the respective scales?  If so, then a given scale (summed or averaged) will contribute to summated score differences across groups more so than will a scale on which there are not variance differences. 

    3. The deeper question is, do responses of the three groups conform to a common factor structure for the set of (you choose: 24 items or 5 scale scores)? 

    4. Both #2 and #3 would be more confidently answered with larger samples than you have at hand.

    Good luck with bringing the debate to a mutually satisfactory conclusion!

    ------------------------------
    David Morse



  • 3.  RE: Advice on the validity of an analysis, please.

    Posted 07-14-2016 08:13

    David,

    Thanks so much for your helpful response.  Here's a short dissertation in reply...

    1. I particularly appreciated your suggestion for adjusting the responses to each subscale by the number of items in the subscale. I was considered using the mean response rather than the summed response for the subscales. Your comment reinforced that it was an appropriate course of action.

    2. Variance: I did a quick and dirty F-test (F=larger subscale variance/smaller subscale variance, ndf=number of responses in group with larger variance minus 1, ddf=number of responses in group with smaller variance minus 1) to test HO: There are no differences in variance. In approximately two-thirds of the combinations, the null hypothesis was rejected so there are significant differences in variance between some subscale responses. Because of this information, I normalized every obervation to the variance and then divided each normalized observation by the number of items in the cognate subscale as suggested in number 1.

    3. Per the factor structure, the authors of the scale did describe the factor structure of the scale. The 24 items were divided into subscales based on the factor structure. As for whether the responses conform to that factor structure, I'm not sure that there are enough responses in each group to begin to determine that. I ran a quick factor analysis of answers to one subscale across all participant groups and the diagnostics suggested that factor analysis is inappropriate.

    In a serendipitous event, your suggestions combined with another member's response in a personal email response resulted in a hallelujah moment and a new approach to addressing this: A factorial analysis. By using this method, I was able to identify that (1) there were no significant interactions between participant groups and subscales, (2) for 4 of the 5 subscales, subscale is the only significant main effect factor that determines the overall scale sum, and (3) for one subscale, both the subscale and the participant group are equally important main effect factors (the interaction was non-significant) that determine the overall scale sum.

    These answers address my hunch; although, not exactly the way in which I thought I would obtain the answer. I think me, myself, and I can stop disagreeing with each other, now.

    Thanks again for your help.  I really appreciate your feedback.

    Linda

    ------------------------------
    Linda A. Landon, PhD, ELS
    Research Communiqué

    Clear, Concise Statistics & Words

    PhD, Molecular Pharmacology
    Graduate Certificate, Applied Statistics
    Board-Certified Editor in the Life Sciences

    Email: LandonPhD@ResearchCommunique.com
    Phone: 573-797-4517
    Website: www.researchcommunique.com
    Jefferson City, MO
    Central Standard Time ( CST ) = GMT-6 (November – February)
    Central Daylight Time ( CDT ) = GMT-5 (March –October)



  • 4.  RE: Advice on the validity of an analysis, please.

    Posted 07-13-2016 11:52

    Here's perhaps another way to approach, though the (relatively) small n is a definite, substantive limitation:

    Use the 5 subscale scores and do multivariate test for differences in the centroids between groups.

    Evaluate the standardized canonical coefficients for the subscales to get an idea of relative importance to group differences.

    Michael

    ------------------------------
    Michael Weaver
    Associate Dean for Research & Scholarship
    University of Florida School of Nursing



  • 5.  RE: Advice on the validity of an analysis, please.

    Posted 07-14-2016 08:29

    Michael,

    Thanks so much for responding to my request.


    Our original plan for analyzing the data from this study was very similar to the approach that you described in your response; we intended to apply canonical correlation and redundancy analysis.  However, when our realized sample size was so low, we altered our analysis plan to accommodate the small sample.

    Our final realized sample size is 36.  My a priori sample size calculation suggested that a minimum sample size of 70 was required to obtain a valid canonical correlation analysis.  Therefore, I think that an overall sample of size 36 is not large enough to produce a valid canonical correlation analysis. Your reply suggests that you would agree with that assessment. Do you have an additional thoughts about that?


    Again, thanks for your help. I do appreciate your input.


    Linda

    ------------------------------
    Linda A. Landon, PhD, ELS
    Research Communiqué

    Clear, Concise Statistics & Words

    PhD, Molecular Pharmacology
    Graduate Certificate, Applied Statistics
    Board-Certified Editor in the Life Sciences

    Email: LandonPhD@ResearchCommunique.com
    Phone: 573-797-4517
    Website: www.researchcommunique.com
    Jefferson City, MO
    Central Standard Time ( CST ) = GMT-6 (November – February)
    Central Daylight Time ( CDT ) = GMT-5 (March –October)