|Back to Essay List||Modified: 23 Jun 2017||BibTeX Entry||RIS Citation|
In general, my topic concerns the “renormalization” of cultural transmission models. This terminology will probably be unfamiliar to anthropologists and social scientists, so I’m not going to emphasize the term or formal renormalization theory in upcoming publications or my dissertation, but it is absolutely what I’m studying. I thought a blog post would be a good place to describe this concept, and its relationship to concepts more familiar to anthropologists.
Those who study long-term records of behavior or evolution face the problem that evolutionary models are individual-based or “microevolutionary,” and describe the detailed change in adoption of traits or flow of genetic information within a population, while our empirical data describe highly aggregated, temporally averaged counts or frequencies. This mismatch in temporal scales is extreme enough that the “evolutionary synthesis” of the 1940′s tended to separate consideration of “microevolution” and “macroevolution” into different sets of processes (largely as a result of George Gaylord Simpson’s pioneering work). The study of the fossil record of life on earth has rightly focused mainly on the phylogenetic history of species and higher taxa, in their paleoecological contexts. When studying the archaeological record of human behavior over shorter (albeit still substantial) time scales, it seems less clear that microevolutionary models cannot inform our explanations.
At the same time, our usage of microevolutionary models of cultural transmission, to date, has almost universally ignored the vast difference in time scales between our evidence and the behavioral events and “system states” we model. The sole exception to this rule, actually, seems to be Fraser Neiman’s 1990 dissertation, which has a sophisticated discussion of the effects of time-averaging on neutral models and cultural trait frequencies. So, an important question would be: what do cultural transmission models look like, when we view their behavior through the lens of a much longer-term data source?
This is precisely the kind of question that renormalization theory answers, as formulated in physics. Below the jump, I describe renormalization in more detail.
Renormalization theory arises in physics whenever we seek to figure out the consequences of a detailed theory at a larger, longer, or slower scale than the theory naturally describes. Renormalization has two origins, first in quantum field theory where it was initially used to remove infinite quantities that prevented finite calculations for the interaction of electrons and photons, and independently within statistical physics in calculating the bulk properties of materials given detailed models of molecular interaction. Kenneth Wilson, in the early 1970′s, unified these two perspectives into the “renormalization group,” for which he later received a Nobel prize.
The basic procedure works like this. Imagine that we want to calculate the force between two largish atoms — say, two iron atoms. The microscopic theory would have us add up the following quantities (at a minimum, assuming both atoms could be treated as stationary): (a) the forces between the two nuclei, (b) the coulomb force between each of the 52-odd electrons and the “other” nucleus, and ( c) the forces between all of the electrons themselves. This is a complicated summation since the electrons are small and “fast” compared to the nuclei, and the force depends upon the distance between the electron and the object to which it is being compared. So what we do in renormalization is recognize that the nuclei are slow and heavy compared to the electrons, and the “net force” between the atoms is really the force between the nuclei with a factor which represents the “average” of the forces exerted by the electrons and between the electrons. The end result is a much simpler average formula which neglects variation on certain time and distance scales, but is hopefully accurate on larger and longer distance and time scales. This is renormalization in a nutshell (the example, by the way, is explained in detail by Leonard Susskind in his terrific series of theoretical physics lectures available on iTunes).
The application of this to cultural transmission theory is fairly straightforward, at least conceptually. We tend to model the dynamics of social learning in single populations, over short-term time scales, and solve for the equilibrium states in those populations. But we observe evidence of that learning, at least outside the laboratory or in ethnographic settings, on much longer time scales. And often in spatially aggregated ways, such that we’re taking samples of the outcome of transmission events over whole communities or areas. Thus, what we need to do is “integrate out” the fast and short term fluctuations and patterns, and see the longer-term swings in trait frequencies and changes in spatial patterns.
This implies a research agenda, since this is several very different tasks. The first is simply understanding the effects of observing cultural transmission models through the lens of temporally and spatially aggregated observations. I made a start on this problem in a conference paper earlier this year, available in preprint form.
The second is recognizing that we nearly always observe social learning in archaeological contexts in regional contexts, where whole communities represent a single sample of the frequencies of cultural variants. This implies a renormalization from single population to metapopulation models, or continuous spatial models. Some work along these lines is underway, and my own contribution is analysis of metapopulation rather than continuous diffusion models (Blythe is also working on this problem).
The third, which is unique to my research as far as I can tell, is not treating our observable data as “snapshots” in time, which is the common pattern in archaeological CT studies. Archaeological data, following from the first point, are temporal aggregates as is well known. Which means that our analysis of transmission models must not compare synchronic or equilibrium predictions to diachronic observations. We must, instead, analyze the dynamics of our models through diachronic, aggregated predictions. This has not yet been done, at least in any literature I’ve seen.
Finally, we observe the archaeological record through formal classifications of artifacts, and do not observe the socially transmitted cultural variation directly. Thus, we cannot directly apply statistical distributions which are designed to apply to DNA sequences or other “more direct” measures of heritable information, to the frequencies of archaeological classes. This problem, of course, exists within population genetics itself given the “allele/locus” models common in the pre-genomic era. But the solution here must be uniquely archaeological given our methods of classification and observation. We must, in other words, view cultural transmission models not through the “traits” we usually model, but analytical classifications that mimic the structure of the multidimensional types and classes we actually use.
I do not, in this research, actually use the formal apparatus of the “renormalization group” from physics. Exploring that apparatus and its applicability would take me far afield from the concrete and useful contributions outlined above. But RG theory and methods are never far from my mind, and after the dissertation is completed, I want to explore its utility in more detail.