13  Structural Equation Modelling

13.1 Introduction

Structural Equation Modelling (SEM) combines factor analysis and multiple regression to analyse complex relationships between variables. SEM allows us to test theoretical models by examining both direct and indirect relationships between observed (measured) variables and unobserved (latent) constructs.

Why is SEM useful?

It’s particularly valuable when multiple interrelated variables need to be analysed simultaneously. It enables us to:

  • Test complex theoretical frameworks;

  • Account for measurement error in analyses;

  • Examine both direct and indirect effects; and

  • Compare alternative models using goodness-of-fit indices.

While more complex than some traditional methods, SEM provides a robust framework for testing hypotheses about relationships between variables and for evaluating theoretical models against empirical data.

Path analysis and SEM

You may have heard the term “path analysis” before, and wondered about the relationship between path analysis and SEM.

“Path analysis” is a precursor to, and fundamental component of SEM.

While path analysis works with observed variables only, SEM extends this by incorporating latent variables (theoretical constructs) and measurement models. This evolution marked a significant advancement in statistical modelling, allowing the testing of more complex theoretical relationships.

Visualisation of a Path Analysis - note that it only includes observed variables, not latent variables

“Observed” variables are things we can directly measure in some way. “Latent” variables are theoretical constructs we cannot directly measure but infer from multiple observed variables.

Examples include:

  • Depression – not measured directly, but assessed through multiple symptoms and behavioural markers

  • Intelligence – not measured directly, but assessed through various test scores

  • Job satisfaction – not measured directly, but inferred from multiple questions about workplace experiences

13.2 Latent and Observed Variables

Introduction

Structural Equation Modelling (SEM) allows us to analyse relationships between different types of variables.

In SEM, there are two categories of variable:

  1. Latent variables: Theoretical constructs that cannot be directly measured but are inferred from multiple observed variables.

  2. Observed variables: Measurable indicators that represent directly collected data. They may act as proxies for latent variables or appear as standalone variables in the model.

Notice how this idea is similar to factor analysis, which also assumes there are latent factors underpinning observed responses.

In SEM diagrams, latent variables are often depicted as circles, while observed variables are represented as rectangles. Arrows indicate relationships: one-headed arrows typically show causal paths, and two-headed arrows indicate correlations.

Latent Variables and SEM

Introduction

Latent variables, often called “unobserved” or “hidden” variables, are a defining feature of SEM. They represent abstract concepts (e.g., intelligence, anxiety, market sentiment) inferred from observed data.

Role of latent variables in SEM

Latent variables form a conceptual bridge between theory and data. They help account for measurement error by allowing a distinction between the underlying construct and the imperfect indicators that measure it. This makes SEM particularly useful in fields where the phenomena of interest are inherently unobservable (e.g., psychological traits, sociological constructs).

Conceptual foundations

The assumption underlying latent variables is that multiple observed indicators reflect a shared underlying construct. For example, if we hypothesize a latent variable “job satisfaction,” we might measure it using several survey items (e.g., pay satisfaction, work environment, teamwork). Each indicator captures a facet of the broader theoretical concept.

Estimation and Identification

Latent variables and their relationships to observed indicators are expressed through equations in SEM. Parameters (factor loadings, structural coefficients) are often estimated using maximum likelihood (ML) or Bayesian methods. A key challenge is model identification. A model is identified if there is enough information in the data to estimate unique parameter values. Common rules of thumb include having at least three good indicators for each latent variable.

Role in structural relationships

In SEM, latent variables can act as predictors and outcomes, forming complex networks of causal relationships. For instance, “academic motivation” could predict “academic performance,” which then predicts “career satisfaction.” SEM can also model indirect effects, such as “parental support” influencing “academic performance” through “self-efficacy.”

Some practical considerations

Because latent variables are not directly measurable, strong theoretical justification and carefully chosen indicators are essential. Validity and reliability must be assessed (e.g., through model fit indices, reliability measures) to ensure that the latent variables accurately capture the constructs.

Measurement models

In SEM, measurement models specify how latent variables are linked to observed indicators. This is often done via confirmatory factor analysis (CFA). Good measurement models ensure that each observed indicator provides useful information about the underlying latent construct.

Observed Variables

Introduction

Observed variables (or “manifest” variables) are directly measured data points—such as test scores, questionnaire items, or physiological readings. They anchor theoretical constructs to empirical reality.

Observed variables in SEM

Observed variables play three main roles in SEM:

  1. Indicators of latent variables

    • E.g., questionnaire items about mood or energy levels could indicate “depression.”
  2. Exogenous variables (independent variables)

    • E.g., parental income or hours spent studying might predict latent constructs or other observed measures.
  3. Endogenous variables (dependent variables)

    • E.g., test scores influenced by latent or observed predictors.

Theoretical implications of observed variables

Choosing the right observed variables is crucial. They must align with the constructs and hypotheses under investigation. For instance, if the theoretical model suggests that “emotional intelligence” predicts “job performance,” the observed indicators chosen (e.g., questionnaire items for EI, supervisor ratings for performance) must be valid and reliable representations of those constructs.

Indicator reliability

Indicator reliability refers to how consistently an observed variable measures what it is intended to measure. In SEM, reliability can be evaluated in several ways:

  • Cronbach’s alpha or composite reliability for sets of indicators

  • Factor loadings in the measurement model (indicators should load strongly on the latent factor)

  • Residual variances to see how much unexplained variation remains

High reliability is critical: if indicators do not consistently reflect the underlying construct, estimates of relationships between variables can be biased or unstable.

13.3 Model Specification in SEM

Introduction

Model specification in SEM involves translating a theoretical framework into a structured set of relationships among variables. Good specification accurately reflects the hypotheses and ensures testable predictions.

What is model specification?

Model specification is the process of defining:

  1. Variables

    • Latent variables (unobservable) and observed variables (directly measurable).
  2. Relationships

    • One-headed arrows to imply causation, two-headed arrows to imply correlation.
  3. Constraints

    • Which parameters are free to be estimated (e.g., path coefficients) and which are fixed.

Measurement models and structural models

  1. Measurement Model

    • Links latent variables to their observed indicators.

    • Commonly specified through Confirmatory Factor Analysis (CFA).

  2. Structural Model

    • Describes how latent (or sometimes observed) variables relate to each other.

    • E.g., “Team Cohesion” → “Match Performance.”

Steps in model specification

  1. Develop a Theoretical Framework

    • Base your model on existing theory or research.
  2. Create a Path Diagram

    • Use circles for latent variables and rectangles for observed variables.
  3. Write the Equations

    • Each path corresponds to an equation (e.g., Performance = β1*(Cohesion) + ε).
  4. Specify Parameters

    • Decide which paths are estimated and which are fixed at 0 or at a specific value.
  5. Check Identifiability

    • Ensure there are enough constraints and data to estimate the model’s parameters uniquely.

Example: Model specification in rugby

Imagine you’re studying factors that influence Team Performance in rugby. You hypothesise:

  • Team Cohesion improves Team Performance

  • Leadership Quality enhances Team Cohesion

Observed variables could include:

  • Survey items for Team Cohesion (e.g., “Trust among players,” “Shared goals”).

  • Coach ratings and player feedback for Leadership Quality.

  • Match outcomes or points scored for Team Performance.

Model specification means defining these latent variables, linking them to observed indicators, and laying out the causal paths.

This is lavaan 0.6-20
lavaan is FREE software! Please report any bugs.
Warning: lavaan->lav_object_post_check():  
   some estimated ov variances are negative
lavaan 0.6-20 ended normally after 40 iterations

  Estimator                                         ML
  Optimization method                           NLMINB
  Number of model parameters                        21

  Number of observations                           200

Model Test User Model:
                                                      
  Test statistic                                34.271
  Degrees of freedom                                24
  P-value (Chi-square)                           0.080

Model Test Baseline Model:

  Test statistic                              2013.860
  Degrees of freedom                                36
  P-value                                        0.000

User Model versus Baseline Model:

  Comparative Fit Index (CFI)                    0.995
  Tucker-Lewis Index (TLI)                       0.992

Loglikelihood and Information Criteria:

  Loglikelihood user model (H0)              -1704.437
  Loglikelihood unrestricted model (H1)      -1687.301
                                                      
  Akaike (AIC)                                3450.874
  Bayesian (BIC)                              3520.139
  Sample-size adjusted Bayesian (SABIC)       3453.608

Root Mean Square Error of Approximation:

  RMSEA                                          0.046
  90 Percent confidence interval - lower         0.000
  90 Percent confidence interval - upper         0.079
  P-value H_0: RMSEA <= 0.050                    0.537
  P-value H_0: RMSEA >= 0.080                    0.044

Standardized Root Mean Square Residual:

  SRMR                                           0.039

Parameter Estimates:

  Standard errors                             Standard
  Information                                 Expected
  Information saturated (h1) model          Structured

Latent Variables:
                   Estimate  Std.Err  z-value  P(>|z|)   Std.lv  Std.all
  Leadership =~                                                         
    LQ1               1.000                               0.946    1.005
    LQ2               0.975    0.042   23.272    0.000    0.922    0.877
    LQ3               0.974    0.041   23.853    0.000    0.922    0.883
  Cohesion =~                                                           
    TC1               1.000                               1.024    1.002
    TC2               0.966    0.039   24.543    0.000    0.989    0.886
    TC3               1.003    0.039   25.570    0.000    1.027    0.895
  Perf =~                                                               
    TP1               1.000                               1.073    0.997
    TP2               0.986    0.034   29.086    0.000    1.059    0.917
    TP3               1.005    0.038   26.665    0.000    1.078    0.900

Regressions:
                   Estimate  Std.Err  z-value  P(>|z|)   Std.lv  Std.all
  Cohesion ~                                                            
    Leadership       -0.052    0.076   -0.688    0.491   -0.048   -0.048
  Perf ~                                                                
    Leadership        0.223    0.072    3.098    0.002    0.197    0.197
    Cohesion          0.424    0.067    6.326    0.000    0.405    0.405

Variances:
                   Estimate  Std.Err  z-value  P(>|z|)   Std.lv  Std.all
   .LQ1              -0.009    0.018   -0.528    0.597   -0.009   -0.010
   .LQ2               0.255    0.030    8.383    0.000    0.255    0.231
   .LQ3               0.240    0.029    8.228    0.000    0.240    0.220
   .TC1              -0.004    0.018   -0.237    0.813   -0.004   -0.004
   .TC2               0.269    0.032    8.472    0.000    0.269    0.216
   .TC3               0.263    0.032    8.220    0.000    0.263    0.200
   .TP1               0.006    0.017    0.345    0.730    0.006    0.005
   .TP2               0.213    0.027    7.925    0.000    0.213    0.159
   .TP3               0.273    0.032    8.489    0.000    0.273    0.190
    Leadership        0.894    0.090    9.913    0.000    1.000    1.000
   .Cohesion          1.046    0.106    9.893    0.000    0.998    0.998
   .Perf              0.928    0.095    9.793    0.000    0.805    0.805

13.4 Parameter Estimation in SEM

Introduction

After specifying a model, the next step is to estimate its parameters (e.g., factor loadings, path coefficients). These estimates tell us how strongly variables are related and help us evaluate whether our theoretical model aligns with the data.

Model parameters

Model parameters in SEM typically include:

  • Factor loadings: how strongly each observed indicator relates to its latent variable.

  • Regression weights (path coefficients): the strength of direct effects between variables.

  • Covariances or correlations: the relationships between exogenous variables or error terms.

  • Variances: how much each variable varies, including error or residual variance.

Estimation techniques

Common estimation methods include:

  • Maximum Likelihood (ML): The most widely used. Assumes multivariate normality and aims to find parameter values that maximise the likelihood of observing the data.

  • Weighted Least Squares (WLS): Often used for categorical or non-normal data. Minimises weighted squared differences between observed and model-implied correlations.

  • Bayesian Estimation: Incorporates prior distributions for parameters, updates these with observed data, and produces posterior distributions.

The choice of estimator depends on the data structure (e.g., continuous vs. ordinal, normal vs. non-normal) and the research question.

Goodness-of-fit

Goodness-of-fit tells us how well the estimated model reproduces the observed data. While specific fit indices appear in the next section, an overall principle is to ensure that the model’s predicted relationships closely match the actual relationships in the dataset. Poor fit often indicates that key paths or constructs are missing, or that the theoretical model is not supported by the data.

13.5 Model Fit and Evaluation

Introduction

Once parameters are estimated, we need to assess how well the model fits the data. Fit indices help determine whether the theoretical model is plausible or needs revision.

Global fit indices

Common global fit indices include:

  • Chi-square test (χ²): Tests the null hypothesis that the model’s implied covariance matrix equals the observed covariance matrix. Sensitive to sample size.

  • Root Mean Square Error of Approximation (RMSEA): Assesses how well the model would fit the population’s covariance matrix. Values ≤ .06 are often considered good.

  • Comparative Fit Index (CFI) and Tucker-Lewis Index (TLI): Compare the specified model to a baseline (null) model. Values close to 1 indicate better fit.

  • Standardized Root Mean Square Residual (SRMR): The average discrepancy between observed and predicted correlations.

Assessment of ‘local fit’

Local fit involves scrutinising individual parameters and residuals:

  • Residuals: Differences between observed and predicted correlations or covariances. Large residuals suggest misfit in specific parts of the model.

  • Modification indices: Indicate how much model fit would improve by freeing a constrained (fixed) parameter. Can guide model refinement, though it’s best to have theoretical justification for any changes.

Model parsimony

Parsimony implies using the simplest model that adequately explains the data. Overly complex models may fit well but risk overfitting, while overly simple models might miss important pathways. Common measures, like the Akaike Information Criterion (AIC), balance fit and complexity.

13.6 Model Validation and Generalisation

Introduction

Even if a model fits well in one dataset, we want to ensure it generalises to other samples and contexts. Validation procedures help confirm the robustness of the findings.

Cross-Validation

Cross-validation involves testing the model on a separate dataset or on different subsets of the same dataset (e.g., split-half validation). If the model fits similarly in new data, it has stronger support for generalisability.

Measurement Invariance

Measurement invariance assesses whether the same latent variables are measured equivalently across groups (e.g., by gender or culture). If invariance holds, group comparisons of latent variable means and relationships are more valid.

Replication

Replication involves re-running the same model with new data or in a different context. Repeated successful replications increase confidence that the findings are not sample-specific and truly reflect underlying relationships.