• We are sorry, but NCBI web applications do not support your browser and may not function properly. More information
Logo of jepicomhInstructions for authorsCurrent TOCJournal of Epidemiology and Community Health
J Epidemiol Community Health. Nov 2007; 61(11): 933–937.
PMCID: PMC2465602

A simple guide to chaos and complexity

Abstract

The concepts of complexity and chaos are being invoked with increasing frequency in the health sciences literature. However, the concepts underpinning these concepts are foreign to many health scientists and there is some looseness in how they have been translated from their origins in mathematics and physics, which is leading to confusion and error in their application. Nonetheless, used carefully, “complexity science” has the potential to invigorate many areas of health science and may lead to important practical outcomes; but if it is to do so, we need the discipline that comes from a proper and responsible usage of its concepts. Hopefully, this glossary will go some way towards achieving that objective.

Keywords: nonlinear dynamics, chaos theory, complexity

The concepts of complexity and chaos are being increasingly invoked in the health sciences literature (general treatments are outlined in references1,2,3). Applications to date include (there are many more):

However, despite their being so widespread, the concepts underpinning complex systems science and chaos theory are still foreign to many health scientists and there is some looseness in how they have been translated from their origins in mathematics and physics, which is leading to much confusion and error in their application.36 This glossary attempts to resolve these issues by providing a simple (but not simplified) guide to many central concepts in chaos theory and complexity science. The many references provide more detail.

Dynamical systems

System is simply the name given to an object studied in some field and might be abstract or concrete; elementary or composite; linear or nonlinear; simple or complicated; complex or chaotic. Complex systems are highly composite ones, built up from very large numbers of mutually interacting subunits (that are often composites themselves) whose repeated interactions result in rich, collective behaviour that feeds back into the behaviour of the individual parts. Chaotic systems can have very few interacting subunits, but they interact in such a way as to produce very intricate dynamics. Simple systems have very few parts that behave according to very simple laws. Complicated systems can have very many parts too, but they play specific functional roles and are guided by very simple rules. Complex systems can survive the removal of parts by adapting to the change; to be robust, other systems must build redundancy into the system (eg by containing multiple copies of a part). A large healthcare system will be robust to the removal of a single nurse because the rest of the members of the system will adapt to compensate—however, adding more nurses does not necessarily make the system more efficient.23,24,37 On the other hand, a complicated piece of medical technology, such as a positron emission tomography scanner, will obviously not survive removal of a major component. The behaviour of a chaotic system appears random, but is generated by simple, non‐random, deterministic processes: the complexity is in the dynamical evolution (the way the system changes over time driven by numerous iterations of some very simple rule), rather than the system itself.59,60

Systems possess properties that are represented by variables or observables: quantities that have a range of possible values such as the number of people in a population, the blood pressure of an individual, the length of time between consultations, and so on. The values taken by a system's variables at an instant of time describe the system's state. A state is often represented by a point in a geometrical space (phase space), with axes corresponding to the variables. The coordinates then correspond to particular assignments of values to each variable. The number of variables defines the dimension of both the space and the system. Each point in the phase space represents a way in which the system could be at a time, corresponding to an assignment of particular values to the variables at an instant. The overall health state of an individual, for example, might include values for lung capacity, heart rate, blood‐sugar levels and so on. A path through the phase space corresponds to a trajectory of the system, or a way in which the system could evolve over time— for example, the change in a person's health state over time (itself a function of many other variables).

A dynamical system is a system whose state (and variables) evolve over time, doing so according to some rule. How a system evolves over time depends both on this rule and on its initial conditions—that is, the system's state at some initial time. Feeding this initial state into the rules generates a solution (a trajectory through phase space), which explains how the system will change over time; chaos is generated by feeding solutions back into the rule as a new initial condition. In this way, it is possible to say what state the system will be in at a particular time in the future (Abraham and Shaw 61 offer an exceptionally clear, graphical introduction to many aspects of dynamical systems theory, including chaos).

Complex and chaotic systems are both examples of nonlinear dynamical systems. A linear system is characterised by the satisfaction of the superposition principle. The superposition principle says that if A and B are both solutions for some system (ways in which the system could evolve), then so is their sum A + B—this implies that a linear system can be decomposed into its parts and each part solved separately to construct the full solution. For nonlinear systems, this is not possible because of the appearance of nonlinear terms, functions of the variables such as sin(x), x3 and xy. In this sense then, the whole here is more than the sum of its parts. Given this, a nonlinear system is one for which inputs are not proportional to outputs: a small (large) change in some variable or family of variables will not necessarily result in a small (large) change in the system. This kind of behaviour is well known to those involved in intervention research: large interventions, in some variable, do not necessarily have a large effect on some outcome variable of interest. Likewise, a small intervention can have large, unexpected outcomes.24,25

A system (or process) is deterministic if it is possible to uniquely determine its past and future trajectories (ie all points it passed through and will pass through) from its initial (present) state. A system is semi‐deterministic if its future but not its past trajectory can be uniquely determined. An indeterministic system is one without a unique future trajectory, so that the evolution is random. It is often possible to tell whether or not a system is deterministic by inspecting the time series it generates, plotting states at different times. The data points in the time series might correspond to measurements of blood sugar levels, population health indicators and so on. If closely matched points at are found in different places in the series, then the series is investigated to see if the points that follow are closely matched too. Hence, attempts are made to infer from the spread of points the kind of system or process that generated it.

The time‐series can also be used to indicate whether or not a system is chaotic or complex: a chaotic system's time‐series has a fractal‐like structure, meaning that it looks the same at different scales (a property also known as self‐similarity: take a snapshot of the time series covering a certain interval of time and then take another snapshot covering a much larger period of time and the two will look the same). Fractals are really just a spatial version of chaos; the interesting type of chaos is the temporal kind. Inferring complexity from a time‐series is more difficult but can be done (this involves finding power law correlations in the data; see below).

A related concept is that of the attractor.45 Following an intervention in a system (changing the value of some variable), it takes a little while for it to settle down into its “normal” behaviour. The path traced out in phase space during this period is known as a transient (or the start‐up transient). The state it reaches after this corresponds to the normal behaviour of the system: the phase space points corresponding to this form the system's attractor. If the attractor is a point that does not move, it is known as a fixed point. Such attractors often describe dissipative systems (those that lose energy— for example, due to friction). In general, however, it is possible to think of an attractor as whatever the system behaves like after it has passed the transient stage.

There are other types of attractor. For example, an attractor that describes a system that cycles periodically over the same set of states, never coming to rest, is known as a limit cycle. A system need not possess a single attractor either; the phase space can have a number of attractors whose “attractiveness” depends upon the initial conditions of the system (ie the state of the system at the outset).

The set of points that are “pulled” towards a particular attractor are known as the basin of attraction. What is important to note about the types of attractor discussed above (fixed points and limit cycles) is that initial points that are close to each other remain close as they each evolve according to the same rules. This property is violated for so‐called strange attractors for which close points diverge exponentially over time. Such attractors are also aperiodic; systems described by strange attractors do not wind up in a steady state nor do they repeat the same pattern of behaviour (an example that can be easily observed in the household is a dripping water tap tuned to a certain flow, not too high amd not too low). Chaotic systems have strange attractors; complex systems have evolving phase spaces and a range of possible attractors.

These concepts have been applied extensively, accurately and successfully in the biomedical sciences.42,47,48 The general outcome of these investigations appears to be that chaos is associated with “good health”: pathologies (such as of the brain, heart, lungs) occur when the dynamics becomes stable and the attractor is a limit cycle.42,46,49,50 Heart disease, epilepsy, bipolarity and so on are considered to be dynamic diseases in that they are not associated with something that can manifest itself at an instant (like a broken arm), but only arise over time.43 This is why chaos is so appropriate.

Chaos versus complexity

We can now consider further the similarities and differences between chaotic systems and complex systems. Each shares common features, but the two concepts are very different. Chaos is the generation of complicated, aperiodic, seemingly random behaviour from the iteration of a simple rule. This complicatedness is not complex in the sense of complex systems science, but rather it is chaotic in a very precise mathematical sense. Complexity is the generation of rich, collective dynamical behaviour from simple interactions between large numbers of subunits. Chaotic systems are not necessarily complex, and complex systems are not necessarily chaotic (although they can be for some values of the variables or control parameter; see below).62

The interactions between the subunits of a complex system determine (or generate) properties in the unit system that cannot be reduced to the subunits (and that cannot be readily deduced from the subunits and their interactions). Such properties are known as emergent properties. In this way it is possible to have an upward (or generative) hierarchy of such levels, in which one level of organisation determines the level above it, and that level then determines the features of the level above it.59 Emergent properties may also be universal or multiply realisable in the sense that there are many diverse ways in which the same emergent property can be generated. For example, temperature is multiply realisable: many configurations of the same substance can generate the same temperature, and many different types of substance can generate the same temperature. The properties of a complex system are multiply realisable since they satisfy universal laws—that is, they have universal properties that are independent of the microscopic details of the system. Emergent properties are neither identical with nor reducible to the lower‐level properties of the subunits because there are many ways for emergent properties to be produced.

A necessary condition, owing to nonlinearity, of both chaos and complexity is sensitivity to initial conditions. This means that two states that are very close together initially and that operate under the same simple rules will nevertheless follow very different trajectories over time. This sensitivity makes it difficult to predict the evolution of a system, as this requires the initial state of the system to be described with perfect accuracy. There will always be some error in how this is performed and it is this error that gets exponentially worse over time. It is possible to see how this might pose problems for replication of initial conditions in various types of trial and intervention.24,25

There are several less well‐understood, but nonetheless important properties that are characteristic features of complex systems. Complex systems often exhibit self‐organisation, which happens when systems spontaneously order themselves (generally in an optimal or more stable way) without “external” tuning of a control parameter (see below). This feature is not found in chaotic systems and is often called anti‐chaos.46 Such systems also tend to be out of equilibrium, which means that the system never settles in to a steady state of behaviour. This is related to the concept of openness: a system is open if it is not or cannot be screened off from its environment. In closed systems, outside influences (exogenous variables) can be ignored. For open systems, this is not the case. Most real‐world systems are open, thus this presents problems both for modelling and experimenting on such systems, because the effect of exogenous influences must be taken into account. Such influences can be magnified over time by sensitivity to initial conditions.

Another important feature of a complex system is the idea of feedback, in which the output of some process within the system is “recycled” and becomes a new input for the system. Feedback can be positive or negative: negative feedback works by reversing the direction of change of some variable; positive feedback increases the rate of change of the variable in a certain direction. In complex systems, feedback occurs between levels of organisation, micro and macro, so that the micro‐level interactions between the subunits generate some pattern in the macro‐level that then “back‐reacts” onto the subunits, causing them to generate a new pattern, which back‐reacts again and so on. This kind of “global to local” positive feedback is called coevolution, a term originating in evolutionary biology to describe the way organisms create their environment and are in turn moulded by that environment.

If a system is stable under small changes in its variables, so that it does not change radically when interventions occur then it is said to be robust. Generally, complex systems increase in robustness over time because of their ability to organise themselves relative to their environment. However, it is possible for single events to alter a complex system in a way that persists for a long time (this is called path‐dependence). For a complex system, “history matters”. Complex processes (processes generated by complex systems) are, therefore, non‐Markovian: they have a long “memory”. This non‐Markovian aspect (essentially due to feedback mechanisms) is often believed to pose insuperable problems for describing causal events and making causal predictions about complex systems. If this were so, it would be a severe blow to the practical usefulness of complex science; however, there has been some encouraging preliminary work involving non‐Markovian and modified Markovian causal models and networks.63,64

The key differences between chaotic systems and complex ones lie, therefore, in the number of interacting parts and the effect that this has on the properties and behaviour of the system as a whole. As Nobel Laureate Phillip Anderson put it: “more is different”.65 Complex systems are coherent units in a way that chaotic systems are not, involving instead interactions between units. This simple difference concerning units and subunits can be brought out using concepts from the theory of critical phenomena, which is central to complexity science.

Critical phenomena

Some systems have a property known as criticality. A system is critical if its state changes dramatically given some small input. To make sense of this, we need to introduce several new concepts.

An order parameter is a macroscopic (global or systemic) feature of the system that tells one how the parts of the system are competing or cooperating with one another. Hence, there is a state of order among the parts when they act collectively (in which case the order parameter is non‐zero) and a state of disorder when they fight against each other, doing the opposite of their neighbours (in which case the order parameter is zero).

The control parameter is an external input to the system that can be varied so as to change the order parameter and so the macroscopic features of the system. This is termed tuning the control parameter to shift the system between various phases or regimes; it is possible to have ordered, chaotic and critical (edge of chaos) phases. A common example concerns the phenomenon of magnetism in a piece of metal. Here the order parameter is the degree of magnetism and the control parameter is temperature. As the magnet is heated up, the magnetism decreases; increasing the control parameter decreases the order parameter. The system undergoes a phasetransition so that at a critical temperature the magnetism vanishes. This is a general feature of complex systems; tuning the system's control parameter to a certain critical point results in a phase transition at which the system undergoes an instantaneous radical change in its qualitative features (the phases of water, from gas to liquid to solid, is another common example). The general study of such behaviour is the theory of critical phenomena.

A system that is at a critical point has an extremely high degree of connectivity between its subunits: everything depends on everything else! Complex systems are said to be poised at such a position, between order and chaos. The degree of connectedness is encoded in the correlation function. This function tells us by how much pairs of subunits are influencing one another and how much this influence varies with distance. The furthest distance the influence extends is known as the correlation length; beyond this distance, the subunits are independent and are unaffected by one another. The farther apart two subunits are, the less they influence each other (the influence effect decreases exponentially with distance). The correlation length can itself be influenced by the control parameter. When the control parameter is tuned to the critical point, the correlation length becomes infinite, and all the subunits follow each other. The influence still decays exponentially with distance, but in the critical regime, more pathways are opened up between pairs of subunits so that the connectivity of the system is massively amplified. As a consequence, a small disturbance in the system (even to a single subunit) can produce massive, systemic changes. Very different kinds of critical system exhibit the same properties—for example, crowds behaving like the atoms in a magnet. This feature is known as universality.14

Universality is connected to another concept: scaling. Scaling laws, or power laws have the following form: f(x) ~ x−α in other words, the probability f(x) of an event of magnitude x occurring is inversely proportional to x.12 Sociologist Vilfred Pareto noticed that the statistical distribution for the wealth of individuals in a population followed such a law: few are rich, some are well‐off and many are poor. Whenever systems are described by scaling laws that share the same exponents (the α term) then they will exhibit similar behaviour in some way (hence, universality). Such systems are said to belong to the same universality class.66 For example, diverse countries follow Pareto's law, as do cities. This result could easily be extrapolated to the distribution of the health of individuals within and across populations, with significant implications for research on health inequalities.

If a system displays power law behaviour then it is scale‐free and its parts have scale‐invariant correlations between them. What this means is that the system does not possess a characteristic scale in the sense that events of all magnitudes can occur. To make sense of this, take an example of a system that does have a characteristic scale: human height. Most humans are of about the same height. There are a few “outliers” who are either very tall or very small, but most are around 5′ 8″. Now consider earthquakes. These can be both extremely tiny and extremely massive, but most are negligible. This is an example of a system that does not have a characteristic scale; it satisfies a scaling law (the Gutenberg–Richter law). Note that self‐similarity , scale invariance and power laws are just equivalent ways of saying that there is no characteristic scale. Various empirical studies have confirmed that healthcare systems obey power laws for a number of quantities of interest, such as hospital waiting lists.4,16,23,24,37 These studies have clear policy implications; if various healthcare delivery systems exhibit power‐law behaviour then we ought to reframe our intervention and management theories in terms of complex systems science. We should not be surprised if huge catastrophes occur for no discernible reason, and we should not be surprised if our massive intervention to reduce waiting times by employing more staff does nothing for efficiency or even makes it worse. However, we should show caution in using the power‐law behaviour to infer an underlying complex system, as power laws can be generated in diverse ways.67,68

Conclusion

Although complexity science is still a work in progress, having neither a firm mathematical foundation nor the necessary and sufficient conditions whose satisfaction entails that some system is complex, it nonetheless has the potential to invigorate many areas of health research as we hope to have indicated in this brief guide. However, if it is to achieve its potential we need the discipline that comes from careful and proper usage of its concepts. Hopefully, this glossary will go some way towards achieving that objective.

What this paper adds

This glossary is intended to provide a “corrective” to what the authors perceive to be a loose and misleading handling of ideas stemming from the sciences of complexity. The aim is to provide as simple an account as possible, without succumbing to the “popular science”‐style accounts that appear to be fast becoming the norm in the health‐sciences literature. The going is perhaps a little tough, but we feel the payoff (a responsible guide) is worth the extra effort.

Footnotes

Funding: This work was conducted as part of an International Collaboration on Complex Interventions (ICCI) funded by the Canadian Institutes of Health Research. DR is an ICCI Post Doctoral Fellow. AS and PH are Senior Scholars of the Alberta Heritage Foundation for Medical Research. PH holds the Markin Chair in Health and Society at the University of Calgary.

References

1. Sweeney K, Griffiths F. eds. Complexity and healthcare: an introduction. Radcliffe Medical Press, 2002
2. Lindberg C, Plsek P E, Zimmerman B. Edgeware: insights from complexity science for health care leaders. Dallas: VHA Inc, 1998
3. Plsek P E, Greenhalgh T. The challenge of complexity in healthcare. BMJ 2001. 323625–628.628 [PMC free article] [PubMed]
4. Philippe P, Mansi O. Nonlinearity in the epidemiology of complex health and disease processes. Theor Med Bioeth 1998. 19591–607.607 [PubMed]
5. Simmons M. Complexity theory in the management of communicable diseases. J Hosp Infec 2003. 5487–92.92 [PubMed]
6. Cohen I, Segel L A. Design principles for the immune system and other distributed autonomous systems. Oxford: Oxford University Press, 2001
7. Philippe P. Epidemiology and self‐organized critical systems: An analysis in waiting times and disease heterogeneity. Nonlinear Dynamics Psychol Life Sci 2000. 4275–295.295
8. Anderson R M. Populations, infectious disease and immunity: a very nonlinear world. Phil Trans Soc Lond B 1994. 346457–505.505 [PubMed]
9. Philippe P. Chaos and public health: Implications for an epidemic. Can J Public Health 1992. 83165–166.166 [PubMed]
10. Philippe P. Chaos, population biology and epidemiology: some research implications. Hum Biol 1993. 65525–546.546 [PubMed]
11. Trottier H, Philippe P. Scaling properties of childhood infectious diseases epidemics before and after mass vaccination in Canada. J Theor Biol 2005. 235326–337.337 [PubMed]
12. Rhodes C J, Anderson R M. Power laws governing epidemics in isolated populations. Nature 1996. 381600–602.602 [PubMed]
13. Philippe P. The scale‐invariant spatial clustering of leukemia in San Francisco. J Theor Biol 1999. 199371–381.381 [PubMed]
14. Pastor‐Satorras R, Vespignani A. Epidemic dynamics and endemic states in complex networks. Phys Rev E 2001. 63066117 [PubMed]
15. Fraser S W, Greenhalgh T. Coping with complexity: Educating for capability. BMJ 2001. 323799–803.803 [PMC free article] [PubMed]
16. Papadopoulos M C, Hadjitheodossiou M, Chrysostomu M. et al Is the National Health Service at the edge of chaos? J R Soc Med 2001. 94613–616.616 [PMC free article] [PubMed]
17. Plsek P E. Re‐designing healthcare with insights from the science of complex adaptive systems. Crossing the chasm: a new health system for the 21st century. Natl Acad Sci 2000. 322355
18. Plsek P E, Wilson T. Complexity science: complexity, leadership and management in healthcare organisations. BMJ 2001. 323746–749.749 [PMC free article] [PubMed]
19. Mark A, Critten P. Chaos and complexity ‐ the future for health care HRM. Health Manpow Manage 1998. 24139–142.142 [PubMed]
20. Haigh C. Utilising chaos theory in the maintenance of nursing services. APS Publishing 2003
21. Dreibe D, McDaniel R. Complexity science and healthcare management. In Blair JD, Fottler MD, Savage GT (eds.), Advances in health care management , vol 2. Stamford, CN: JAI Press 2001. 11–36.36
22. Byng R, Jones R. Mental health link: The development and formative evaluation of a complex intervention to improve shared care for patients with long‐term mental illness. J Eval Clin Pract 2004. 1027–36.36 [PubMed]
23. Smethurst D P, Williams H C. Are hospital waiting lists self‐regulating? Nature 2001. 410652–653.653 [PubMed]
24. Philippe P, Garcia M R, West B J. Evidence of “essential uncertainty” in emergency‐ward length of stay. Fractals 2004. 12197–209.209
25. Fullilove R E, Edgoose J C, Fullilove M T. Chaos, criticality and public health. J Nat Med Assoc 1997. 89311–316.316 [PMC free article] [PubMed]
26. Kernick D. ed. Complexity and healthcare organization: A view from the street. Oxford: Radcliffe Medical Press 2004
27. Sweeney K, Griffiths F. Complexity: a new model for post‐normal medicine. J Eval Clin Pract 2002. 18356–358.358
28. Wilson T, Holt T, Greenhalgh T. Complexity and clinical care. BMJ 2001. 323685–688.688 [PMC free article] [PubMed]
29. Martin C M, Sturmberg J P. General practice—chaos, complexity and innovation. MJA 2005. 183106–109.109 [PubMed]
30. Miller M. Crabtree B, McDaniel R. Understanding change in primary care practice using complexity theory. J Fam Pract 1998. 46369–376.376 [PubMed]
31. Wilson T, Holt T, Greenhalgh T. Complexity and clinical care. BMJ 2001. 323685–688.688 [PMC free article] [PubMed]
32. Heath I, Sweeney K. Complexity in primary care: understanding its value. Oxford: Radcliffe Medical Press, 2006
33. Griffiths F, Byrne D. General practice and the new science emerging from the theories of “chaos” and complexity. Br J Gen Pract 1998. 481697–1699.1699 [PMC free article] [PubMed]
34. Innes A D, Campion P D, Griffiths F E. Complex consulations and the “Edge of Chaos”. Br J Gen Pract 2005. 5547–52.52 [PMC free article] [PubMed]
35. Kernick D. Wanted—new methodology for health services research. Is complexity theory the answer? Fam Pract 2006. 23385–390.390 [PubMed]
36. Marshall M. Chaos and complexity. Br J Gen Pract 1999. 49234 [PMC free article] [PubMed]
37. Love T, Burton C. General practice as a complex system: a novel analysis of consultation data. Fam Pract 2005. 22347–352.352 [PubMed]
38. Holt T. ed. Complexity for clinicians. Oxford: Radcliffe Medical Press 2004
39. Steinberg D. Complexity in healthcare and the language of consultation: exploring the other side of medicine. Oxford: Radcliffe Medical Press 2005
40. Bassingthwaighte J B, Liebovitch L S, West B J. Fractal physiology. Oxford: Oxford University Press, 1994
41. Babloyantz A, Destexhe A. Low‐dimensional chaos in an instance of epilepsy. Proc Natl Acad Sci U S A 1986. 833513–3517.3517 [PMC free article] [PubMed]
42. Otero‐Siliceo E, Arriada‐Mendicoa N. Is it healthy to be chaotic? Med Hypotheses 2003. 60233–236.236 [PubMed]
43. Glass L, Mackey M C. From clocks to chaos: The rhythms of life. p. 1998.
44. Suki B, Barabasi A ‐ L, Hantos Z. et al Avalanches and power‐law behaviour in lung inflation. Nature 1994. 368615–618.618 [PubMed]
45. Denton T A, Diamond G A, Helfant R H. et al Fascinating rhythm: a primer on chaos theory and its application to cardiology. Am Heart J 1990. 1201419–1440.1440 [PubMed]
46. Coffey D S. Self‐organization, complexity and chaos: The new biology for medicine. Nature 1998. 4882–885.885 [PubMed]
47. Deisboeck T S, Yasha Kresh J. eds. Complex systems science in biomedicine. New York: Springer 2006
48. Quarteroni A, Formaggia L, Veneziani A. eds. Complex systems in biomedicine. New York: Springer 2006
49. Glass L, Kaplan D. Complex dynamics in physiology and medicine. In: Weigend AS, Gershenfeld NA, eds. Time series prediction: Forecasting the future and understanding the past. Santa Fe Institute Studies in the Sciences of Complexity: Addison Wesley, 1994. 513–528.528
50. Lipsitz L A, Goldberger A L. Loss of “Ccomplexity” and aging. Potential applications of fractals and chaos theory to senescence. JAMA 1992. 141806–1809.1809 [PubMed]
51. Philippe P, West B J. The complex dynamics of diabetes modeled as a fractal complex‐adaptive‐system. Rev Bras Epidemiol 1998. 1280–293.293
52. Sataloff R, Hawkshaw M. Chaos and medicine: source readings. San Diego: Singular Press, 2001
53. Walleczek J. Self‐organized biological dynamics and nonlinear control: toward understanding complexity, chaos and emergent function in living systems. Cambridge: Cambridge University Press, 2000
54. Goldberger A L, West B J. Chaos in physiology: health or disease? In: Holteon A, Olsen LF, eds. Chaos in biological systems. New York: Plenum Press, 1987. 1–5.5
55. Albrecht G, Freeman S, Higginbotham N. Complexity and human health: the case for a transdisciplinary paradigm. Cult Med Psychiatry 1998. 2255–92.92 [PubMed]
56. Higginbotham N, Albrecht G, Connor L. Health social science: A transdisciplinary and complexity perspective. Oxford: Oxford University Press, 2001
57. Gatrell A C. Complexity theory and geographies of health: A critical assessment. Soc Sci Med 2004. 602661–2671.2671 [PubMed]
58. Blackman T. Placing health: neighbourhood renewal, health improvement and complexity. Bristol: Policy Press, 2006
59. May R M. Simple mathematical models with very complicated dynamics. Nature 1976. 261459–467.467 [PubMed]
60. May R M. Simple rules with complex dynamics. Science 2000. 287601–602.602 [PubMed]
61. Abraham R H, Shaw C D. Dynamics: the geometry of behaviour. Redwood City, CA: Addison‐Wesley Publishing Company, 1992
62. Bertuglia C S, Vaio F. Nonlinearity, chaos and complexity: The dynamics of natural and social systems. Oxford: Oxford University Press, 2005
63. Haslberger A, Varga F, Karlic H. Recursive causality in evolution: A model for epigenetic mechanisms in cancer development. Med Hypotheses 2006. 621448–1454.1454 [PubMed]
64. Williamson J, Gabbay D. Recursive causality in Bayesian networks and self‐fibring networks. In: Gillies D, ed. Laws and models in science. London: King”s College Publications, 2005. 173–221.221
65. Anderson P W. More is different. Science 1972. 177393–96 [PubMed]
66. Cvitanovic P. ed. Universality in chaos, 2nd ed. Bristol: IOP Publishing Ltd 1989
67. Frigg R. Self‐organised criticality — what it is and what it isn't. Stud Hist Phil Sci 2002. 34613–632.632
68. Newman M E J. Power laws, pareto distributions and Zipf”s law. Contemporary Physics 2005. 46323–351.351

Articles from Journal of Epidemiology and Community Health are provided here courtesy of BMJ Group
PubReader format: click here to try

Formats:

Related citations in PubMed

See reviews...See all...

Cited by other articles in PMC

See all...

Links

  • PubMed
    PubMed
    PubMed citations for these articles

Recent Activity

Your browsing activity is empty.

Activity recording is turned off.

Turn recording back on

See more...