Thursday, March 27, 2014

Are we winning the science communication war?

Since the time that I was a young graduate student, there have been constant calls for ecologists to communicate more with the public and policy makers (Norton 1998, Ludwig et al. 2001). The impetus for these calls is easy to understand –we are facing serious threats to the maintenance of biodiversity and ecosystem health, and ecologists have the knowledge and facts that are needed to shape public policy. To some, it is unconscionable that ecologists have not done more advocacy, while others see a need to better educate ecologists in communication strategies. While the reluctance for some ecologists to engage in public communication could be due to a lack of skills that training could overcome, the majority likely has had a deeper unease. Like all academics, ecologists have many demands on their time, but are evaluated by research output. Adding another priority to their already long list of priorities can seem overwhelming. More fundamentally, many ecologists are in the business of expanding our understanding of the world. They see themselves as objective scientists adding to global knowledge. To these ‘objectivists’, getting involved in policy debates, or becoming advocates, undermines their objectivity.

Regardless of these concerns, a number of ecologists have decided that public communication is an important part of their responsibilities. Ecologists now routinely sit on the boards of different organizations, give public lectures, write books and articles for the public, work more on applied problems, and testify before governmental committees. Part of this shift comes from organizations, such as the Nature Conservancy, which have become large, sophisticated entities with communication departments. But, the working academic ecologist likely talks with more journalists and public groups than in the past.

The question remains: has this increased emphasis on communication yielded any changes in public perception or policy decisions. As someone who has spent time in elementary school classrooms teaching kids about pollinators and conservation, the level of environmental awareness in both the educators and children surprises me. More telling are surprising calls for policy shifts from governmental organizations. Here in Canada, morale has been low because of a federal government that has not prioritized science or conservation. However signals from international bodies and the US seem to be promising for the ability of science to positively influence science.

Two such policy calls are extremely telling. Firstly, the North American Free Trade Agreement (NAFTA), which includes the governments of Mexico, Canada, and the USA, which normally deals with economic initiatives and disagreements, announced that they will form a committee to explore measures to protect monarch butterflies. They will consider instituting toxin-free zones, where the spraying of chemicals will be prohibited, as well as the construction of a milkweed corridor from Canada to Mexico. NAFTA made this announcement because of declining monarch numbers and calls from scientists for a coordinated strategy.

The second example is the call from 11 US senators to combat the spread of Asian carp. Asian carp have invaded a number of major rivers in the US, and have their spread has been of major concern to scientists. The 11 senators have taken this scientific concern seriously, requesting federal money and that the Army Corps of Engineers devise a way to stop the Asian carp spread.


There seems to be promising anecdotal evidence that issues of scientific concern are influencing policy decisions. This signals a potential shift; maybe scientists are winning the public perception and policy war. But the war is by no means over. There are still major issues (e.g., climate change) that require more substantial policy action. Scientists, especially those who are effective and engaged, need to continue to communicate with public and policy audiences. Every scientifically informed policy decision should be seen as a signal of the willingness of audiences to listening to scientists and that communicating science can work.



References
Ludwig D., Mangel M. & Haddad B. (2001). ECOLOGY, CONSERVATION, AND PUBLIC POLICY. Annual Review of Ecology and Systematics, 32, 481-517.

Norton, B. G. 1998. IMPROVING ECOLOGICAL COMMUNICATION: THE ROLE OF ECOLOGISTS IN ENVIRONMENTAL POLICY FORMATION. Ecological Applications 8:350–364


Monday, March 24, 2014

Debating the p-value in Ecology

It is interesting that p-values still garner so much ink: it says something about how engrained and yet embattled they are. This month’s Ecology issue explores the p-value problem with a forum of 10 new short papers* on the strengths and weaknesses, defenses and critiques, and various alternatives to “the probability (p) of obtaining a statistic at least as extreme as the observed statistic, given that the null hypothesis is true”.

The defense for p-values is lead by Paul Murtaugh, who provides the opening and closing arguments. Murtaugh, who has written a number of good papers about ecological data analysis and statistics, takes a pragmatic approach to null hypothesis testing and p-values. He argues p-values are not flawed so much as they are regularly and egregiously misused and misinterpreted. In particular, he demonstrates mathematically that alternative approaches to the p-value, particularly the use of confidence intervals or information theoretic criteria (e.g. AIC), simply present the same information as p-values in slightly different fashions. This is a point that the contribution by Perry de Valpine supports, noting that all of these approaches are simply different ways of displaying likelihood ratios and the argument that one is inherently superior ignores their close relationship. In addition, although acknowledging that cutoff values for significant p-values are logically problematic (why is a p-value of 0.049 so much more important than one of 0.051?), Murtaugh notes that cutoffs reflect decisions about acceptable error rates and so are not inherently meaningless. Further, he argues that dAIC cutoffs for model selection are no less arbitrary. 

The remainder of the forum is a back and forth argument about Murtaugh’s particular points and about the merits of the other authors’ chosen methods (Bayesian, information theoretic and other approaches are represented). It’s quite entertaining, and this forum is really a great idea that I hope will feature in future issues. Some of the arguments are philosophical – are p-values really “evidence” and is it possible to “accept” or “reject” a hypothesis using their values? Murtaugh does downplay the well-known problem that p-values summarize the strength of evidence against the null hypothesis, and do not assess the strength of evidence supporting a hypothesis or model. This can make them prone to misinterpretation (most students in intro stats want to say “the p-value supports the alternate hypothesis”) or else interpretation in stilted statements.

Not surprisingly, Murtaugh receives the most flak for defending p-values from researchers working in alternate worldviews like Bayesian and information-theoretic approaches. (His interactions with K. Burnham and D. Anderson (AIC) seem downright testy. Burnham and Anderson in fact start their paper “We were surprised to see a paper defending P values and significance testing at this time in history”...) But having this diversity of authors plays a useful role, in that it highlights that each approach has it's own merits and best applications. Null hypothesis testing with p-values may be most appropriate for testing the effects of treatments on randomized experiments, while AIC values are useful when we are comparing multi-parameter, non-nested models. Bayesian similarly may be more useful to apply to some approaches than others. This focus on the “one true path” to statistics may be behind some of the current problems with p-values: they were used as a standard to help make ecology more rigorous, but the focus on p-values came at the expense of reporting effect sizes, making predictions, or careful experimental design.

Even at this point in history, it seems like there is still lots to say about p-values.

*But not open-access, which is really too bad. 

Monday, March 17, 2014

How are we defining prediction in ecology?

There is an ongoing debate about the role of wolves in altering ecosystem dynamics in Yellowstone, which has stimulated a number of recent papers, and apparently inspired an editorial in Nature. Entitled “An Elegant Chaos”, the editorial reads a bit like an apology for ecology’s failure at prediction, suggesting that we should embrace ecology’s lack of universal laws and recognize that “Ecological complexity, which may seem like an impenetrable thicket of nuance, is also the source of much of our pleasure in nature”.

Most of the time, I also fall squarely into the pessimistic “ecological complexity limits predictability” camp. And concerns about prediction in ecology are widespread and understandable. But there is also something frustrating about the way we so often approach ecological prediction. Statements such as “It would be useful to have broad patterns and commonalities in ecology” feel incomplete. Is it that we really lack “broad patterns and commonalities in ecology”, or has ecology adopted a rather precise and self-excoriating definition for “prediction”? 

.
We are fixated on achieving particular forms of prediction (either robust universal relationships, or else precise and specific numerical outputs), and perhaps we are failing at achieving these. But on the other hand, ecology is relatively successful in understanding and predicting qualitative relationships, especially at large spatial and temporal scales. At the broadest scales, ecologists can predict the relationships between species numbers and area, between precipitation, temperature and habitat type, between habitat types and the traits of species found within, between productivity and the general number of trophic levels supported. Not only do we ignore this foundation of large-scale predictable relationships, but we ignore the fact that prediction is full of tradeoffs. As a paper with the excellent title, “The good, the bad, and the ugly of predictive science” states, any predictive model is still limited by tradeoffs between: “robustness-to-uncertainty, fidelity-to-data, and confidence-in-prediction…. [H]igh-fidelity models cannot…be made robust to uncertainty and lack-of-knowledge. Similarly, equally robust models do not provide consistent predictions, hence reducing confidence-in-prediction. The conclusion of the theoretical investigation is that, in assessing the predictive accuracy of numerical models, one should never focus on a single aspect.” Different types of predictions have different limitations. But sometimes it seems that ecologists want to make predictions in the purest, trade-off free sense - robustness-to-uncertainty, fidelity-to-data, and confidence-in-prediction - all at once. 

In relation to this, ecological processes tend to be easier to represent in a probabilistic fashion, something that we seem rather uncomfortable with. Ecology is predictive in the way medicine is predictive – we understand the important cause and effect relationships, many of the interactions that can occur, and we can even estimate the likelihood of particular outcomes (of smoking causing lung cancer, of warming climate decreasing diversity), but predicting how a human body or ecosystem will change is always inexact. The complexity of multiple independent species, populations, genes, traits, all interacting with similarly changing abiotic conditions makes precise quantitative predictions at small scales of space or time pretty intractable. So maybe that shouldn’t be our bar for success. The analogous problem for an evolutionary biologist would be to predict not only a change in population genetic structure but also the resulting phenotypes, accounting for epigenetics and plasticity too. I think that would be considered unrealistic, so why is that where we place the bar for ecology? 

In part the bar for prediction is set so high because the demand for ecological knowledge, given habitat destruction, climate change, extinction, and a myriad of other changes, is so great. But in attempting to fulfill that need, it may be worth acknowledging that predictions in ecology occur on a hierarchy from those relationships at the broadest scale that we can be most certain about, moving down to the finest scale of interactions and traits and genes where we may be less certain. If we see events as occurring with different probabilities, and our knowledge of those probability distributions declining the farther down that hierarchy we travel, then our predictive ability will decline as well. New and additional research adds to the missing or poor relationships, but at the finest scales, prediction may always be limited.

Tuesday, March 11, 2014

The lifetime of a species: how parasitism is changing Darwin's finches

Sonia Kleindorfer, Jody A. O’Connor, Rachael Y. Dudaniec, Steven A. Myers,Jeremy Robertson, and Frank J. Sulloway. (2014). Species Collapse via Hybridization in Darwin’s Tree Finches. The American Naturalist, Vol. 183, No. 3, pp. 325-341

Small Galapagos tree finch,
Camarhynchus parvulus
Darwin’s finches are some of the best-known examples of how ecological conditions can cause character displacement and even lead to speciation. Continuing research on the Galapagos finches has provided the exceptional opportunity to follow post-speciation communities and explore how changes in ecological processes affect species and the boundaries between them. Separate finch species have been maintained in sympatry on the islands because various barriers maintain the species' integrity, preventing hybrids from occurring (e.g. species' behavioural differences) or being recruited (e.g. low fitness). As conditions change though, hybrids may be a source of increased genetic variance and novel evolutionary trajectories and selection against them may weaken. Though speciation is interesting in its own right, it is not the end of the story: ecological and evolutionary pressures continue and species continue to be lost or added, to adapt, or to lose integrity.

A fascinating paper by Kleindorfer et al. (2014) explores exactly this issue among the small, medium, and large tree finches (Camarhynchus spp.of Floreana Island, Galapagos. Large and small tree finches first colonized Floreana, with the medium tree finch speciating on the island from a morph of the large tree finch. This resulted in three sympatric finch species that differ in body and beak size, but otherwise share very similar behaviour and appearance. However, ecological and environmental conditions have not remained constant on Floreana since observations in the 1800s: a parasite first observed on the island in 1997, Philornis downsi has taken residence and has caused massive nestling mortality (up to 98%) for the tree finches. Since parasite density is correlated with tree finch body size, the authors predicted that high parasite intensity should be linked to declining recruitment of the large tree finch. If females increasingly prefer smaller mates, there may also be increased hybridization, particularly if there is some advantage in having mixed parental ancestry. To test this, the authors sampled tree finch populations on Floreana in both 2005 and 2010. Parasite numbers increase with high precipitation, and so by combining museum records (collected between 1852-1906, when no parasites were present), 2005 sampling records (dry conditions, lower parasite numbers), and 2010 sampling records (high rainfall, high parasite numbers), they could examine a gradient of parasite effects. They measured a number of morphological variables, collected blood for genotyping, estimated individual age, measured parasite intensity in nests, and observed mate choice.

Philornis downsi:
larval stage parasitizes nestlings.
(a Google image search will provide
some more graphic illustrations)
For each time period, morphological measurements were used to cluster individuals into putative species. The museum specimens from the 1800s had 3 morphologically distinguishable populations, the true small, medium and large tree finch species usually written about. In 2005 there were still 3 distinct clusters, but the morphological overlap between them had increased. By 2010, the year with the highest parasite numbers, there were only two morphologically distinguishable populations. Which species had disappeared? Although recent studies have labelled the two populations as the “small” tree finch and “large” tree finch, the authors found that the 2010 “large” population is much smaller than the true large tree finches collected in 1852-1906, suggesting perhaps the large tree finch was no longer present. Genetic population assignment suggested that despite morphological clustering, there were actually only two distinct species on Floreana in 2005 and 2010: it appeared that the large tree finch species had gone extinct, and the boundary between the small and medium tree finch species had become porous, leading to morphologically intermediate hybrids.

The question then, is whether the extinction of the large tree finch and the collapse of the boundary between small and medium tree finches can be attributed to the parasite, and the changing selective pressures associated with it. Certainly there were clear changes in size structure (from larger birds to smaller birds) and in recruitment (from few young hybrids to many young hybrids) between the low parasite year (2005) and the high parasite year (2010). Strikingly, parasite loads in nests were much lower for hybrids and smaller-bodied populations than for the larger-bodied population (figure below). Compared to their large-bodied parents, hybrids somehow avoided parasite attack even in years with high parasite densities (2010). When parasite loads are high, hybrid offspring have a fitness advantage, as evidenced by the large number of young hybrids in 2012. The collapse of the large tree finch population is also likely a product of parasite pressures as well, as females selected smaller mates with comparatively lower parasite loads. Despite the apparent importance of the parasites in 2010, the existence of only a few older hybrid individuals, and greater morphological distance between populations seen in the 2005 survey (a low parasite period) suggests that selection for hybrids varies greatly through time. Though the persistence of the Philornis parasite on Floreana may prevent re-establishment of the large tree finch, changing parasite densities and other selective pressures may continue to cause the boundaries of the remaining finch populations to overlap and retract in the future. The story of Darwin's finches is even more interesting if we consider that it doesn't stop at character displacement but continues to this day.
From Kleindorfer et al 2014: Philornis parasite intensity in nests sampled in 2005 (lower parasite) and 2010/2012 (higher parasite), for nests of the small-bodied (population 1), intermediate hybrid, and larger-bodied (population 2) individuals.

Friday, March 7, 2014

EEB & Flow inclusion in Library of Congress Web Archives

I just received this email the other day, and nearly deleted it as another spam email (along with fake conference invites, obscure journal submission invites, and offers to make millions). But apparently it's legit, and the US Library of Congress has been archiving web sites for some time. They are now building a collection of science blogs (link, link), which is a pretty cool idea, and we're excited to be part of it :-)

Monday, February 24, 2014

Evolution at smaller and smaller scales: a role for microgeographic adaptation in ecology?

Jonathan L. Richardson, Mark C. Urban, Daniel I. Bolnick, David K. Skelly. 2014. Microgeographic adaptation and the spatial scale of evolution. Trends in Ecology & Evolution, 19 February 2014.

Among other trends in ecology, it seems that there is a strong trend towards re-integration of ecological and evolutionary dynamics, and also in partitioning ecological dynamics to finer and finer scales (e.g. intraspecific variation). So it was great to see a new TREE article on “Microgeographic adaptation and the spatial scale of evolution”, which seemed to promise to contribute to both topics.

In this paper, Richardson et al. attempt to define and quantify the importance of small-scale adaptive differences that can arise between even neighbouring populations. These are given the name “microgeographic adaptation”, and defined as arising via trait differences across fine spatial scales, which lead to fitness advantages in an individual’s home sites. The obvious question is what spatial scale does 'microgeographic' refer to, and the authors define it very precisely as “the dispersal neighborhood … of the individuals located within a radius extending two standard deviations from the mean of the dispersal kernel of a species”. (More generally they forward an argument for a unit--the ‘wright’--that would measure adaptive divergence through space relative to dispersal neighbourhoods.) The concept of microgeographic adaptation feels like it is putting a pretty fine point on already existing ideas about local adaptation, and the authors acknowledge that it is a special case of adaptation at scales where gene flow is usually assumed to be high. Though they also suggest that microgeographic adaptation has received almost no recognition, it is probably fairer to say that in practice the assumption is that on fine scales, gene flow is large enough to swamp out local selective differences, but many ecologists could name examples of trait differences between populations at close proximity.

From Richardson et al. (2014). One
example of microgeographic adaptations.
Indeed, despite the general disregard to fine-scale evolutionary differences, they note that there are some historical and more recent examples of microgeographic variation. For example, Robert Selander found that despite the lack of physical barriers to movement, mice in neighbouring barns show allelic differences, probably due to territorial behaviour. As you might expect, microgeographic adaptations result when migration is effectively lower than expected given geographic distance and/or selection is stronger (as when neighbouring locations are very dissimilar). A variety of mechanisms are proposed, including the usual suspects – strong natural selection, landscape barriers, habitat selection, etc.

A list of the possible mechanisms leading to microgeographic adaptation is rather less interesting than questions about how to quantify the importance and commonness of microgeographic adaptation, and especially about its implications for ecological processes. At the moment, there are just a few examples and fewer still studies of the implications, making it difficult to say much. Because of either the lack of existing data and studies or else the paper's attempt to be relevant to both evolutionary biologists and ecologists, the vague discussion of microgeographic differences as a source of genetic variation for restoration or response to climate change, and mention of the existing—but primarily theoretical—ecological literature feels limited and unsatisfying. The optimistic view is that this paper might stimulate a greater focus on (fine) spatial scale in evolutionary biology, bringing evolution and ecology closer in terms of shared focus on spatial scale. For me though, the most interesting questions about focusing on smaller and smaller scales (spatial, unit of diversity (intraspecific, etc)) are always about what they can contribute to our understanding. Does complexity at small scales simply disappear as we aggregate to larger and larger scales (a la macroecology) or does it support greater complexity as we scale up, and so merit our attention? 

Tuesday, February 18, 2014

P-values, the statistic that we love to hate

P-values are an integral part of most scientific analyses, papers, and journals, and yet they come with a hefty list of concerns and criticisms from frequentists and Bayesians alike. An editorial in Nature (by Regina Nuzzo) last week provides a good reminder of some of the more concerning issues with the p-value. In particular, she explores how the obsession with "significance" creates issues with reproducibility and significant but biologically meaningless results.

Ronald Fischer, inventor of the p-value, never intended it to be used as a definitive test of “importance” (however you interpret that word). Instead, it was an informal barometer of whether a test hypothesis was worthy of continued interest and testing. Today though, p-values are often used as the final word on whether a relationship is meaningful or important, on whether the the test or experimental hypothesis has any merit, even on whether the data is publishable. For example in ecology, significance values from a regression or species distribution model are often presented as the results. 

This small but troubling shift away from the original purpose for p-values is tied to concerns about false alarms and with replicability of results. One recent suggestion for increasing replicability is to make p-values more stringent - to require that they be less that 0.005. But the point the author makes is that although p-values are typically interpreted as “the probability of obtaining a test statistic at least as extreme as the one that was actually observed, assuming that the null hypothesis is true”, this doesn't actually mean that a p-value of 0.01 in one study is exactly consistent with a p-value of 0.01 found in another study. P-values are not consistent or comparable across studies because the likelihood that there was a real (experimental) effect to start with alters the likelihood that a low p-value is just a false alarm (figure). The more unlikely the test hypothesis, the more likely a p-value of 0.05 is a false alarm. Data mining in particular will be (unwittingly) sensitive to this kind of problem. Of course one is unlikely to know what the odds of the test hypothesis are, especially a priori, making it even more difficult to correctly think about and use p-values. 

from: http://www.nature.com/news/scientific-method-statistical-errors-1.14700#/b5
The other oft-repeated criticism of p-values is that a highly significant p-value make still be associated with a tiny (and thus possibly meaningless) effect size. The obsession with p-values is particularly strange then, given that the question "how large is the effect?", should be more important than just answering “is it significant?". Ignoring effect sizes leads to a trend of studies showing highly significant results, with arguably meaningless effect sizes. This creates the odd situation that publishing well requires high profile, novel, and strong results – but one of the major tools for identifying these results is flawed. The editorial lists a few suggestions for moving away from the p-value – including to have journals require effect sizes and confidence intervals be included in published papers, to require statements to the effect of “We report how we determined our sample size, all data exclusions (if any), all manipulations and all measures in the study”, in order to limit data-mining, or of course to move to a Bayesian framework, where p-values are near heresy. The best advice though, is quoted from statistician Steven Goodman: “The numbers are where the scientific discussion should start, not end.”

Monday, February 10, 2014

Ecological progress, what are we doing right?

A post from Charles Krebs' blog called "Ten limitations on progress in ecology" popped up a number of times on social media last week. Krebs is a established population ecologist who has been working in the field for a long time, and he suggests some important problems leading to a lack of progress in ecology. These concerns range from lack of jobs and funding for ecologists, to the fracturing of ecology into poorly integrated subfields. Krebs' post is a continuation of the ongoing conversation about limitations and problems in ecology, which has been up for discussion for decades. And as such, I agree with many of the points being made. But it reminded me of something I have been thinking about for a while, which is that it seems much more rare to see ecology’s successes listed. For many ecologists, it is probably easier to come up with the problems and weaknesses, but I think that's more of a cognitive bias than a sign that ecology is inescapably flawed. And that’s unfortunate: recognizing our successes and advances also helps us improve ecology. So what is there to praise about ecology, and what successes we can build on?

Despite Krebs’ concerns about lack of jobs for ecologists, it is worth celebrating how much ecology has grown in numbers and recognition as a discipline. The first ESA annual meeting in 1914 had 307 attendees, recent years’ attendance is somewhere between 3000-4000 ecologists. Ecology is also increasingly diverse. Ecology and Evolutionary Biology departments are now common in big universities, and sometimes replacing Botany and/or Zoology programs. On a more general level, the idea of “ecology” has increasing recognition by the public. Popular press coverage of issues such as biological invasions, honeybee colony collapses, wolves in Yellowstone, and climate change, have at least made the work of ecologists slightly more apparent.

Long-term ecological research is probably more common and more feasible now than it has ever been. There are long-term fragmentation, biodiversity and ecosystem function studies, grants directed at LTER, and a dedicated institute (the National Ecological Observatory Network (NEON)) funded by the NSF for longterm ecological data collection. (Of course, not all long term research sites have had an easy go of things – see the Experimental Lakes Area in Canada).

Another really positive development is that academic publishing is becoming more inclusive – not only are there more reputable open access publishing options for ecologists, the culture is changing to one where data is available online for broad access, rather than privately controlled. Top journals are reinforcing this trend by requiring that data be published in conjunction with publications.

Multi-disciplinary collaboration is more common than ever, both because ecology naturally overlaps with geochemistry, mathematics, physics, physiology, and others, and also because funding agencies are rewarding promising collaborations. For example, I recently saw a talk where dispersal was considered in the context of wind patterns based on meteorological models. It felt like this sort of mechanistic approach provided a much fuller understanding of dispersal than the usual kernel-based model.

Further, though subdisciplines of ecology have at times lost connection with the core knowledge of ecology, some subfields have taken paths that are worth emulating, integrating multiple areas of knowledge, while still making novel contributions to ecology in general. For example, disease ecology is multidisciplinary, integrating ecology, fieldwork, epidemiological models and medicine with reasonable success.

Finally, more than ever, the complexity of ecology is being equalled by available methods. More than ever, the math, the models, the technology, and the computing resources available are sufficient. If you look at papers from ecology’s earliest years, statistics and models were restricted to simple regressions or ANOVAs and differential equations that could be solved by hand. Though there is uncertainty associated with even the most complex model, our ability to model ecological processes is higher than ever. Technology allows us to observe changes in alleles, to reconstruct phylogenetic trees, and to count species too small to even see. If used carefully and with understanding, we have the tools to make and continue making huge advances.

Maybe there are other (better) positive advances that I’ve overlooked, but it seems that – despite claims to the contrary – there are many reasons to think that ecology is a growing, thriving discipline. Not perfect, but successfully growing with the technological, political, and environmental realities.
Ecology may be successfully growing, but it's true that the timing is rough...

Tuesday, February 4, 2014

Competition and mutualism may be closely related: one example from myrmecochory


Robert J. Warren II, Itamar Giladi, Mark A. Bradford 2014. Competition as a mechanism structuring mutualisms. Journal of Ecology. DOI: 10.1111/1365-2745.12203.

As ecologists usually think about them, competition and mutualism are very different types of interactions. Competition has a negative effect on resource availability for a species, while mutualism should have a positive impact on resource availability. Mutualisms involve interactions between two or more species, and as such are biotic in nature. While the typical definition of the fundamental niche includes all (and only) abiotic conditions necessary for a population’s persistence, with the realized niche showing those areas that are suitable once biotic interactions are considered (Pulliam 2000), mutualisms are a reminder that the a niche is not as simple as we hope. Mutualisms may be necessary for a population’s persistence, as in the case of obligate pollinators, and so some biotic interactions might be “fundamental”. More complicated still, species may compete for mutualist partners – plant species for pollinators, for example. If the mutualist partner is considered a resource, mutualism and competition may not be so far apart after all. 

The relation between competition and mutualism is probably most acknowledged in terms of pollinators – patterns of staggered flowering in a plant community arise in part to decrease simultaneous demand for limited pollinator resources. Another possibly fundamental biotic resource is dispersers, which may be necessary for population persistence of some species. In Warren, Giladi, and Bradford (2014), the authors attempt to expand this idea of competition for mutualist partners to ant-mediated seed dispersal or myrmecochory. Myrmecochorous plant species are common in a number of regions of the world. They rely on ant dispersal to move their seeds, helping to increase the distance between parent and offspring (and thus decrease competition), lower seed predation, and introduce seeds to novel habitats. Ant species that disperse these seeds benefit from the high-energy seed attachment (elaiosome) provided by the plant. While myrmecochorous plants are dependent on ants for successful dispersal, most ants do not rely solely on elaiosomes for food; further, there are fewer seed-dispersing ant species than there are ant-requiring plant species. As a result, competition for ants between myrmecochorous species is a reasonable hypothesis. If there is competition for mutualist partners, the predictions are that species either increase their attractiveness as a competitor by making their seeds most attractive, or else decrease the intensity of competition by staggering seed release.

Warren et al. tested this predictions for eastern North American woodland perennials: at least 50 plant species rely on ant dispersal in this region, but a much smaller number of ants actually disperse seeds. This dearth of mutualist partners implies that competition for ant dispersers should be particularly strong. One way to successfully monopolize a mutualist is to ensure that the timing of seed release is coordinated with ant availability and attraction: in fact comparisons between myrmecochorous and non-myrmecochorous plant species suggests that those requiring ants set seed earlier, when ant attraction to seeds is higher (insect prey become more attractive later in the season). To look at competition within myrmecochorous species, the authors as whether seed size (and thereby attractiveness to ants) was staggered through time. Smaller mymecochore seeds should, for example, become available when larger and more attractive seeds are not in competition. This prediction held – small, less attractive seeds were available earlier in the season than the larger, more attractive later seeds. The authors then experimentally tested whether small and large seeds were in competition for ants and differed in their success in attracting them. Using weigh boats secured to the forest floor, the researchers provided either i) only small myrmecochore seeds, ii) only large seeds, or iii) a combination of both seed sizes. Not that surprisingly, the presence of large seeds inhibits the removal of smaller less attractive seeds by as much as 100% (i.e. no small seeds were removed).

The authors do a nice job of showing that species differ in their success in attracting ant dispersers, and species with differing seed attractiveness appear to partition the season in such a way as to maximize their success. Whether or not this likely competition for dispersers extends to impact the species’ spatial distribution or whether species are prevented from co-occurring by competition for mutualists is less clear, and an interesting future direction. The authors also hypothesize that dispersers, rather than pollinators, may drive flowering/seed production in a system, which is an alternative the usual assumption that pollinators, not dispersers are more important drivers of evolution. More generally, the paper is a reminder that, at least for some species, biotic interactions are fundamental to the niche. Or even more likely, that the separation between the determinants of a fundamental and realized niche aren’t so very distinct. And that’s a reminder that has value for many sections of ecology, from species distribution models to invasive species research.

Wednesday, January 29, 2014

Guest post: One way to quantify ecological communities

This is a guest post by Aspen Reese, a graduate student at Duke University, who in addition to studying the role of trophic interactions in driving secondary succession, is interested in how ecological communities are defined. Below she explains one possible way to explicitly define communities, although it's important to note that communities must explicitly be networks for the below calculations.

Because there are so many different ways of defining “community”, it can be hard to know what, exactly, we’re talking about when we use the term. It’s clear, though, that we need to take a close look at our terminology. In her recent post, Caroline Tucker offers a great overview of why this is such an important conversation to have. As she points out, we aren’t always completely forthright in laying out the assumptions underlying the definition used in any given study or subdiscipline. The question remains then: how to function—how to do and how to communicate good research—in the midst of such a terminological muddle?

We don’t need a single, objective definition of community (could we ever agree? And why should we?). What we do need, though, are ways to offer transparent, rigorous definitions of the communities we study. Moreover, we need a transferable system for quantifying these definitions.

One way we might address this need is to borrow a concept from the philosophy of biology, called entification. Entification is a way of quantifying thingness. It allows us to answer the question: how much does my study subject resemble an independent entity? And, more generally, what makes something an entity at all?

Stanley Salthe (1985) gives us a helpful definition: Entities can be defined by their boundaries, degree of integration, and continuity (Salthe also includes scale, but in a very abstract way, so I’ll leave that out for now). What we need, then, is some way to quantify the boundedness, integration, and continuity of any given community. By conceptualizing the community as an ecological network*—with a population of organisms (nodes) and their interactions (edges)—that kind of quantification becomes possible.

Consider the following framework: 

Boundedness
Communities are discontinuous from the environment around them, but how discrete that boundary is varies widely. We can quantify this discreteness by measuring the number of nodes that don’t have interactions outside the system relative to the total number of nodes in the system (Fig. 1a). 

Boundedness = (Total nodes without external edges)/(Total nodes)

Integration
Communities exhibit the interdependence and connections of their parts—i.e. integration. For any given level of complexity (which we can define as the number of constitutive part types, i.e. nodes (McShea 1996)), a system becomes more integrated as the networks and feedback loops between the constitutive part types become denser and the average path length decreases. Therefore, degree of integration can be measured as one minus the average path length (or average distance) between two parts relative to the total number of parts (Fig. 1b).

Integration 1-((Average path length)/(Total nodes))

Continuity
All entities endure, if only for a time. And all entities change, if only due to entropy. The more similar a community is to its historical self, the more continuous it is. Using networks from two time points, a degree of continuity is calculated with a Jaccard index as the total number of interactions unchanged between both times relative to the total number of interactions at both times (Fig. 1c).

Continuity = (Total edges-changed edges)/(Total edges)
Fig 1. The three proposed metrics for describing entities—(A) boundedness, (B) integration, and (C) continuity—and how to calculate them. 

Let’s try this method out on an arctic stream food web (Parker and Huryn 2006). The stream was measured for trophic interactions in June and August of 2002 (Fig. 2). If we exclude detritus and consider the waterfowl as outside the community, we calculate that the stream has a degree of boundeness of 0.79 (i.e. ~80% of its interactions are between species included in the community), a degree of integration of 0.98 (i.e. the average path length is very close to 1), and a degree of continuity of 0.73 (i.e. almost 3/4 of the interactions are constant over the course of the two months). It’s as easy as counting nodes and edges—not too bad! But what does it mean?
Fig. 2: The food web community in an arctic stream over summer 2002. Derived from Parker and Huryn (2006). 

Well, compare the arctic stream to a molecular example. Using a simplified network (Burnell et al. 2005), we can calculate the entification of the cellular respiration pathway (Fig. 3). We find that for the total respiration system, including both the aerobic and anaerobic pathways, boundedness is 0.52 and integration is 0.84. The continuity of the system is likely equal to 1 at most times because both pathways are active, and their makeup is highly conserved. However, if one were to test for the continuity of the system when it switches between the aerobic and the anaerobic pathway, the degree of continuity drops to 0.6.
Fig. 3: The anaerobic and aerobic elements of cellular respiration, one part of a cell’s metabolic pathway. Derived from Burnell et al. (2005)
Contrary to what you might expect, the ecological entity showed greater integration than the molecular pathway. This makes sense, however, since molecular pathways are more linear, which increases the average shortest distance between parts, thereby decreasing continuity. In contrast, the continuity of molecular pathways can be much higher when considered in aggregate. In general, we would expect the boundedness score for ecological entities to be fairly low, but with large variation between systems. The low boundedness score of the molecular pathway is indicative of the fact that we are only exploring a small part of the metabolic pathway and including ubiquitous molecules (e.g. NADH and ATP).

Here are three ways such a system could improve community ecology: First, the process can highlight interesting ecological aspects of the system that aren’t immediately obvious. For example, food webs display much higher integration when parasites are included, and a recent call (Lafferty et al. 2008) to include these organisms highlights how a closer attention to under-recognized parts of a network can drastically change our understanding of a community. Or consider how the recognition that islands, which have clear physical boundaries, may have low boundedness due to their reliance on marine nutrient subsidies (Polis and Hurd 1996) revolutionized how we study them. Second, this methodology can help a researcher find a research-appropriate, cost-effective definition of the study community that also maximizes its degree of entification. A researcher could use sensitivity analyses to determine what effect changing the definition of her community would have on its characterization. Then, when confronted with the criticism that a certain player or interaction was left out of her study design, she could respond with an informed assessment of whether the inclusion of further parts or processes would actually change the character of the system in a quantifiable way. Finally, the formalized process of defining a study system will facilitate useful conversation between researchers, especially those who have used different definitions of communities. It will allow for more informed comparisons between systems that are similar in these parameters or help indicate a priori when systems are expected to differ strongly in their behavior and controls.

Communities, or ecosystems for that matter, aren’t homogeneous; they don’t have clear boundaries; they change drastically over time; we don’t know when they begin or end; and no two are exactly the same (see Gleason 1926). Not only are communities unlike organisms, but it is often unclear whether or not communities or ecosystems are units of existence at all (van Valen 1991). We may never find a single objective definition for what they are. Nevertheless, we work with them every day, and it would certainly be helpful if we could come to terms with their continuous nature. Whatever definition you choose to use in your own research—make it explicit and make it quantifiable. And be willing to discuss it with your peers. It will make your, their, and my research that much better.