Chapter Twenty

Towards a Science of Design

Learning Objectives
  1. Articulate the argument for treating design as an applied science
  2. Describe the evidence hierarchy for design knowledge
  3. Explain how scientific laws, evolved practices, and empirical testing complement each other
  4. Identify open questions and future directions in usability science
  5. Synthesise the textbook's themes into a coherent framework for evidence-based design

Introduction

This textbook has covered a wide range of material: from the millisecond-level processing of the perceptual system to the centuries-old proportions of Palladian architecture, from the controlled precision of Fitts's Law to the messy reality of clinical software usability. This final chapter steps back to ask a broader question: can design be a science? And if so, what kind of science is it? [Simon, 2019; Cross, 2007] The argument of this textbook is that usability is — or can be — an applied science: a discipline that draws on fundamental knowledge about human beings to produce designs that serve human needs effectively Vincenti, 1993. This chapter synthesises the evidence presented throughout the book and makes the case for a cumulative, evidence-based approach to design Paton, 2022.

The Evidence Hierarchy for Design

Not all design knowledge is created equal. The evidence supporting a design decision can range from personal intuition to replicated controlled experiments. A useful framework organises design evidence into levels, analogous to the evidence hierarchies used in medicine.

Level 1: Expert Opinion and Intuition

A designer's judgment, shaped by experience and training. Valuable but subject to individual bias, limited by personal experience, and difficult to replicate or verify.

Level 2: Evolved Practice

Design conventions that have survived a process of cultural selection over extended periods — architectural proportions, typographic conventions, navigation patterns Alexander, 1977. Chapter 9 argued that evolved practice represents accumulated evidence: designs that worked were imitated; designs that failed were abandoned. The evidence is observational and retrospective, but the sample size (centuries of practice) and the selection pressure (buildings must be lived in, books must be read) give it substantial weight.

Level 3: Design Heuristics and Expert Consensus

Principles such as Nielsen's 10 heuristics Nielsen, 1994, Shneiderman's golden rules Shneiderman, 1987, and Norman's design principles Norman, 2016. These represent distilled expert knowledge, informed by research but not derived from a single experiment. They occupy a middle ground between individual judgment and empirical evidence.

Level 4: Observational Studies

Usability tests, field studies, surveys, and analytics data that measure actual user performance and experience. These provide empirical evidence about specific designs in specific contexts. Their limitation is that they describe what happened, not necessarily why, and their results may not generalise beyond the specific conditions studied.

Level 5: Predictive Models

Fitts's Law Fitts, 1954, the KLM Card, 1983, cognitive load theory Sweller, 1988, and other models that predict human performance from first principles. These models are derived from controlled experiments, have been widely replicated, and can be applied to new designs without empirical testing. They provide the most generalisable form of design evidence, albeit for specific aspects of performance (speed, accuracy, memory load) rather than the full breadth of usability.

Level 6: Controlled Experiments

Randomised controlled experiments that isolate the effect of specific design features on specific outcomes. These provide the strongest evidence for causal claims ("Design A is faster than design B because of feature X"). Chapter 18 covered the methodology; the present chapter considers their role in the broader evidence landscape.

Key Principle

No single level of evidence is sufficient for good design. Controlled experiments provide the strongest causal evidence but cannot cover every design decision. Predictive models generalise well but address only specific aspects of performance. Heuristics provide broad coverage but lack specificity. Evolved practice offers time-tested guidance but may not apply to novel contexts. The strongest design decisions are supported by evidence from multiple levels — a convergence of experiment, model, heuristic, and practice.

The Scientific Laws of Design

Throughout this textbook, we have encountered several quantitative laws that predict human performance:

Fitts's Law (Chapter 5)

MT = a + b × log2(D/W + 1) Predicts pointing time as a function of distance and target width Fitts, 1954. Validated across dozens of input devices and thousands of experiments MacKenzie, 1992. The most robust quantitative law in HCI.

Hick's Law (Chapter 4)

RT = a + b × log2(n + 1) Predicts choice reaction time as a function of the number of equally probable alternatives [Hick, 1952; Hyman, 1953]. Applies to decision among known options; does not apply to visual search for unknown targets.

Miller's Law (Chapter 3)

Working memory capacity ≈ 4 ± 1 chunks for novel information Cowan, 2001 Sets the limit on how many independent items a user can hold in working memory simultaneously. Cowan's revision Cowan, 2001 of Miller's original 7 ± 2 Miller, 1956 is the figure now taken as authoritative. The most important constraint on information display design.

The Power Law of Practice (Chapter 6)

Tn = T1 × n^(-α) Predicts how performance improves with practice. Applies to virtually all skill acquisition, from typing to menu navigation.

The Steering Law (Chapter 5)

T = a + b × (D/W) Predicts movement time through a constrained path (tunnel) Accot, 1997. Governs the difficulty of cascading menus and narrow drag paths.

Weber's Law (Chapter 2)

ΔI / I = constant The just-noticeable difference in a stimulus is a constant proportion of the stimulus intensity. Governs contrast sensitivity and the perception of differences.

Design Law

These laws are not rules of thumb or guidelines — they are empirically validated mathematical relationships with high predictive accuracy. They derive from the fixed architecture of the human perceptual, cognitive, and motor systems. Because human physiology changes slowly (on evolutionary timescales), these laws remain valid as technology changes. Fitts's Law predicted touchscreen performance decades before touchscreens were common, because the law describes the human motor system, not the input device.

The Complementarity of Evidence Sources

The strongest design arguments draw on multiple evidence sources that converge on the same conclusion.

Example

The case for larger touch targets:

  • Fitts's Law (predictive model) Fitts, 1954: mathematically predicts that larger targets are acquired faster
  • Empirical studies (controlled experiments) MacKenzie, 1992: numerous studies confirm faster acquisition and fewer errors with larger targets
  • Usability testing (observational) Rubin, 2008: users observed struggling with small targets on touchscreens
  • Evolved practice (design convention): platform guidelines (Apple HIG Inc., 2024, Material Design) converge on 44–48pt minimum sizes
  • Design heuristics (expert consensus) Nielsen, 1994: Nielsen's "error prevention" and "flexibility and efficiency of use" support adequate target sizing Five independent evidence sources all point to the same conclusion. This convergence makes the design decision robust — it would hold up even if one evidence source were challenged.

Open Questions

Despite the substantial body of knowledge covered in this textbook, many important questions remain open.

Individual Differences

The laws and models in this textbook describe "average" human performance. But individual differences in perception, cognition, and motor skill are substantial. How should designs accommodate this variability? Adaptive interfaces that adjust to individual users offer one approach, but they introduce their own usability challenges (unpredictability, loss of consistency).

Cultural Variation

Most usability research has been conducted in Western, educated, industrialised, rich, democratic (WEIRD) populations. Design conventions that seem universal — reading left to right, navigating top to bottom, using red for danger — are culturally specific. How well do the principles in this textbook generalise across cultures?

Emotional and Aesthetic Dimensions

The laws and models covered here primarily address performance (speed, accuracy, cognitive load). But usability also encompasses satisfaction, trust, engagement, and delight — dimensions that are harder to quantify and predict. Vitruvius's venustas (beauty) is acknowledged but not yet well integrated into the predictive framework.

The Changing Nature of Interaction

The traditional model of usability — a user operating deterministic controls to accomplish defined tasks — is being disrupted by AI (Chapter 19) [Shneiderman, 2022; Norman, 2023]. Conversational interfaces, autonomous agents, and generative systems create interaction patterns for which existing models were not designed. New models, methods, and principles will be needed.

Sustainability and Ethics

Usability has traditionally focused on making designs work well for their users. But design also affects people who are not the direct users — communities affected by data collection, workers displaced by automation, environments degraded by electronic waste. A complete science of design must eventually account for these broader impacts.

Think About It

This textbook presents usability as a science with quantitative laws and predictive models. But design is also a practice — a craft involving judgment, creativity, and values that cannot be reduced to formulas. Is there a tension between the scientific and the craft dimensions of design? Can they be reconciled, or will design always require both the rigour of the scientist and the sensibility of the artist?

A Framework for Evidence-Based Design

Drawing together the threads of this textbook, we can articulate a framework for evidence-based design:

  1. Start with the science. Use the quantitative laws (Fitts's Law, Hick's Law, Miller's Law, cognitive load theory) to establish the fundamental constraints on the design. These laws identify what is physically and cognitively possible.
  2. Apply design heuristics. Use Nielsen's heuristics, Shneiderman's golden rules, and Norman's principles to guide the thousands of design decisions that cannot each be individually modelled. Heuristics provide broad coverage and catch common patterns of failure.
  3. Draw on evolved practice. Consult the design traditions of architecture, typography, and other mature disciplines for guidance on proportion, hierarchy, rhythm, and organisation. These evolved practices represent accumulated evidence from centuries of human experience.
  4. Predict with models. Use GOMS, KLM, Fitts's Law calculations, and cognitive load estimation to make quantitative predictions about specific design alternatives. Models are most useful for comparing options and identifying bottlenecks.
  5. Evaluate with experts. Conduct heuristic evaluations and cognitive walkthroughs to catch problems before involving users. Expert review is fast and inexpensive; use it early and often.
  6. Test with users. Conduct usability testing to discover problems that models and experts miss. Observe real users performing real tasks. There is no substitute for this empirical grounding.
  7. Measure at scale. Use A/B testing and analytics to measure the impact of design changes on real user behaviour at scale. Quantitative data from large populations provides the most definitive evidence of a design's effectiveness.
  8. Iterate. No design is finished. Each round of evidence — from models, experts, users, and analytics — reveals problems and opportunities that feed the next iteration.

Conclusion

The designed world — our buildings, our software, our tools, our cities — mediates our relationship with each other and with the tasks of daily life. When this mediation works well, we accomplish our goals efficiently and comfortably. When it fails, we struggle, err, and sometimes come to harm. The science of usability offers a path to better design: not a guarantee of perfection, but a systematic approach to understanding human capabilities, predicting human performance, and evaluating designed artefacts against human needs. The laws, models, heuristics, and methods presented in this textbook are tools for that work. The tools are imperfect and incomplete. The laws cover specific aspects of performance but not the full richness of human experience. The models predict expert behaviour but not the full range of human variability. The heuristics provide guidance but not certainty. And the methods reveal problems but do not automatically solve them. What the science does provide is a foundation — a body of knowledge, accumulated through experiment and practice, that makes informed design decisions possible. The architect who knows the proportions of Palladio, the interface designer who knows Fitts's Law, the clinician who understands cognitive load — all are better equipped to create designs that serve human beings well. The ambition of a science of design is to make this knowledge cumulative: each study, each validated model, each refined heuristic adds to a growing body of evidence that future designers can draw upon. This textbook is a contribution to that accumulation — an attempt to gather, organise, and make accessible the scientific knowledge that can make designed objects more usable, more efficient, and more humane.

Key Takeaways

  • Design knowledge exists on a hierarchy from expert intuition to controlled experiments. The strongest decisions draw on multiple levels.
  • The scientific laws of usability (Fitts's, Hick's, Miller's, steering, power law of practice, Weber's) derive from human physiology and remain valid as technology changes.
  • Evidence-based design integrates scientific laws, design heuristics, evolved practices, predictive models, expert evaluation, user testing, and large-scale measurement.
  • Open questions include individual differences, cultural variation, emotional dimensions, AI interaction, and ethical implications.
  • The ambition of usability science is cumulative knowledge: each study and model contributes to a growing evidence base for better design.

Further Reading

  • Simon, H. A. (1969). The Sciences of the Artificial (3rd ed., 1996). MIT Press.
  • Card, S. K., Moran, T. P., & Newell, A. (1983). The Psychology of Human-Computer Interaction. Lawrence Erlbaum Associates.
  • Cross, N. (2006). Designerly Ways of Knowing. Springer.
  • Paton, C. (2022). Scientific Human-Centred Design. DPhil Thesis, University of Oxford.
  • Alexander, C. (1964). Notes on the Synthesis of Form. Harvard University Press.
  • Vincenti, W. G. (1990). What Engineers Know and How They Know It. Johns Hopkins University Press.