Annual report of my intuition about the brain (2023)

How does the brain work and how can we understand it? To view this big question from a broad perspective, I’m reporting some ideas about the brain that marked me most during the past twelve months and that, on the other hand, do not overlap with my own research focus. Enjoy the read! And check out previous year-end write-ups: 20182019202020212022, 2023, 2024.

Introduction

During the past year, I have been starting my own junior group at the University of Zurich and spent some time finishing postdoc projects. After a laborious revision process with several challenging experiments, our study on hippocampal astrocytes and how they integrate calcium signals, which was my main postdoc project with Frijtof Helmchen, is now accepted for publication (in principle), and I’m looking forward to venturing into new projects.

For this year-end write up, I want to discuss the book “The unfolding of language” by Guy Deutscher, which I read this Summer. The book is primarily about linguistics and the evolution of languages, but at the end of this blog post, I will connect some of its ideas with current lines of research in systems neuroscience.

Languages as a pathway to understand thinking

After high school, my main interest was to understand how thoughts work. Among the various approaches to this broad question, I felt that I needed to choose between at least two paths. First, to study the physiology of neuronal systems, in order to understand how neurons connect, how they form memories, concepts, ideas, and finally thoughts. Second, to study the matrix that shapes our thoughts and that itself is formed by the dynamics and connections in our brain: language. I decided for the first trajectory, but I always remained curious about how to use language to understand our brain.

I like languages, especially from the theoretical perspective. I was a big fan of Latin in high school and particularly enjoyed to dive into the structure of its language, that is, its grammar. My mother tongue, German, also offered a lot of interesting and complex grammar to explore. And when I read through the books of J.R.R. Tolkien on Middle-earth, I was fascinated by all the world-building but in particular by the languages that he had invented. And I filled a few notebooks with some (less-refined) attempts to create my own languages. Apart from the beautiful words and the intriguing sounds, I was especially into the structure and grammar of these languages. – Later, I became relatively proficient in a few computer languages like C, Python, LaTeX, HTML or Matlab. I briefly tried to learn Chinese for a year, but although I was fascinated by the logical grammatical edifice, I did not have the same energy to internalize the look-up-tables of characters and sounds. Occupied with neuroscience and mostly reading and writing scientific English (a poor rip-off of proper English), I lost touch with the idea of using language to understand our thoughts and, therefore, the brain.

“The unfolding of language”

Then, Mid of last year, I came across the book The unfolding of language in a small but excellent bookstore at Zurich main station. I was captivated immediately. And I enjoyed the book for several reasons.

First, the book is abundant with examples how language evolves, about the origin of words or grammatical structures. Many details that I never had bothered to consider, made suddenly sense, like some old German words, the French conjugation of verbs, or a certain declination of a Latin word. In addition, I learned a few surprising new principles about other languages, for example, the structure underlying the semitic languages, for which the author is an expert.

Second, the depiction of the evolution of language revealed a systematic aspect to these changes, even across languages even from different language families. I did have some basic knowledge about the language sound changes that had been observed around 1800 and described by Grimm’s law. However, Guy Deutscher’s book did not describe these systematic changes as artifacts of our history but as a natural process that occurred like this or similarly in many different languages, independently from each other. To some extent, reading about all these systematic and – to some extent – inevitable changes of language made me more relaxed about the distortions that language undergoes when used by younger generations or people who mix in loanwords from other languages without hesitation; language is a non-static, evolving object. But what is the evolution of language described by Guy Deutscher actually?

The three principles of the evolution of language

Deutscher mentions three main principles underlying the evolution of language. The strong impression that this book left on me was not only through these principles of language evolution, but even more so through the wealth of examples across languages, types of words and grammars that it provided.

First principle: Economy

“Economy”, according to Deutscher, mostly originates from the laziness of the language users, resulting in the omission or merging of words. One of the examples given in the book was the slow erosion of the name of the month “august” from the Latin word augustus over the Old French word aost to août (which is pronounced as /ut/) in modern French. Such erosion of language appeared as the most striking feature of the evolution of language when it was discovered two centuries ago. It was an intriguing observation that the grammar of old languages like Latin or Sanskrit seemed to be so much more complex than the grammar of newer languages like German, Spanish or English. This observation led to the idea that language is not actually simply evolving but rather decaying.

Deutscher, however, describes how this apparent decay can also be the source and driving force for the creation of new structures. One particular compelling example is his description of how the French conjugation for the future tense has evolved. It is difficult to convey this complicated idea in brief terms, but it evolves around the idea that a late Latin expression like amare habeo (“I have to love”) evolved its meaning (to “I will love”), which was transferred to French, where the futur tense therefore contains the conjugation of the present tense of the word “to have”:

j’aimeraii will lovej’aii have
tu aimerasyou will lovetu asyou have
il aimerahe will loveil ahe has
nous aimeronswe will lovenous avonswe have
vous aimerezyou will lovevous avezyou have
ils aimerontthey will loveils ontthey have

I knew both Latin and French quite well, but this connection struck me as both surprising and compelling. Deutscher also comes up with other examples of how grammar was generated by erosion, but you will have to read the book to make up your own mind.

Second principle: Expressiveness

Expressiveness comes from the desire of language users to highlight and stress what they want to say, in order to overcome the natural inflation of meaning through usage. A typical example are the simple words “yes” and “no”, which are simple and short and therefore often enhanced for emphasis (“yes, of course!” or “not at all!”).

A funny example given by Deutscher is the French word aujourd’hui, which means “today” and is one of the first words that an early beginner learns about French. Deutscher points out that this word was derived from the Latin expression hoc die (“on this day”), which eroded to the Old French word hui (“today”). To more strongly emphasize the word, people started to say au jour d’hui, which basically means “on the day of this day”. Later, au jour d’hui was eroded to aujourd’hui. Nowadays, French speakers start using au jour d’aujourd’hui to put more emphasis on the expression. Therefore, the expression means “today” but literally can be decoded as “on the day of the day of this day”. This example illustrates the close interaction of the expressiveness principle and the erosion principle. And it shows that we are carrying these multiple layers of eroded expressiveness with our us, often without noticing ourselves.

Third principle: Analogy

“Analogy” occurs when humans observe irregularities of language and try to impose rules in order to get rid of exceptions that do not really fit in. For example, children might say “the ship sinked” instead of “the ship sank”. Through erosion, language can take a shape that does not make any sense (because its history of evolution, which would shine a light on the shape, is not obvious), and we try to counteract and impose some structure.

Metaphors are the connection between the physical world and abstraction

But there is one more ingredient that, according to Deutscher, drives the evolution of language. It is this aspect which I found most interesting and most closely connected to neuroscience: metaphors. This idea might sound surprising at first, but once it unfolds by means of examples, it becomes more and more convincing. Deutscher depicts metaphors as a way – actually, the way – how the meaning of words can become more abstract over time.

He gives examples of daily used language and then dissects the used words as having roots in the concrete world. These roots had been lost by usage but can still be seen through the layers of erosion and inflation of meaning. For instance, “abstraction” as a word comes from the Latin words ab and strahere, which basically means “to pull something off of sth.”, that is, to remove a word from its concrete meaning. “Pulling sth off”, on the other hand, is something very concrete, rooted in the physical word.

Such abstraction of meanings is most obvious for loanwords from other languages (here, from Latin). But Deutscher brings up convincing examples of how this process occurs as well for words that evolved within a given language. To give a very simple example that also highlights what Deutscher means when he speaks of metaphors: in the expressions “harsh measures”, the “harsh” has roots in the physical world, describing for example the roughness of a surface (“rough” or, originally, “hairy”). Later, however, “harsh” has been applied to abstract concepts such as “measures”– originally as a metaphor that we, however, do no longer perceive as such. Deutscher recounts many more examples, which, in their simplicity, are sometimes quite eye-opening. He makes the fascinating point that all abstracts words are rooted in the physical world and are therefore mostly dead metaphors. And how could one not agree with this hypothesis? Because, what else can be the origin of a word if not physical reality?

How metaphors create the grammar of languages

Deutscher, however, even goes beyond this idea and posits that abstraction and metaphors may have created also more complex aspects of language. For example, in most languages, there are three perspectives: me, you and they. He makes the point that these perspectives might have derived from demonstratives. “Here” transforms to “me”, “there” to “you”, and a third word for something more distant “over there” to “they”. All of these words are “pointing” words, probably deriving from and originally accompanying pointing gestures. In English, the third kind of word does not really exist as clearly, but for example Japanese features the threefold distinction between koko (“here”), soko (“there”) and asoko (“over there”). The third category is represented in Latin by the word ille, which refers to somebody who is more distant. As a nice connection, the Latin word ille was the origin for the French word il/ils, which means “he/they”. This shows how the metaphorical use of words related to the physical world (point words) can generate abstract concepts like grammar, here: the third person. Deutscher alos brings up languages where the connection between the three demonstratives for persons of variable distances and the pronouns for me/you/they is more directly visible, for example in Vietnamese.

Therefore, Deutscher plausibly demonstrates how not only abstract words, but also more complex structures that underlie the most basic grammar, are evolved from metaphorical usage of concrete words that are related to the physical world, either because they describe physical things (the surface roughness) or because they are originally only enhancements of gestures (pointing words). These ideas in the book are amongst the most interesting ones that I have encountered for quite some time.

Embedding of “The unfolding of language” in current research

Overall, I like the hypotheses presented by Deutscher. The only issue I have is that there are many missing bits and pieces that prevent me from properly see through all potential weaknesses. I simply don’t know whether these questions are unanswered in general or whether Deutscher did not have enough space to treat them in this (popular science) book. Put differently, the book was very engaging and a fascinating read but did not provide useful links to the research literature. How are all these ideas embedded in current linguistics research, or is all of this Deutscher’s own concepts? I observed that there are some links in this work to ideas like the conceptual metaphor or linguistic relativity, but I was unable to really figure out where to get started if I wanted to dig deeper into the principal role of metaphorical use for the development of grammar and abstraction. If a linguistics expert somehow happens to read this blog post, I’d be really happy to get a recommendation on a standard textbook (if there is any) on these topics and how Guy Deutscher’s work fits in.

Metaphors, abstraction and system neuroscience

However, I’d like to briefly discuss an aspect of the metaphor principle that I found particularly interesting, also because of a potential, albeit loose, link to current research in neuroscience. Many neuroscientists are probably aware of the large branch of neuroscience dedicated to understanding the generation and representation of abstract knowledge. This can take very different forms, but one of the most prominent takes is the idea that the hippocampus and the entorhinal cortex, originally shown to represent space (via the famous place cells and grid cells, respectively), also represents more abstract knowledge.

The entorhinal cortex represents physical space using the above-mentioned grid cells that span space with hexagonal lattices of varying spatial scales. Building upon the hexagonal lattice idea, researchers attempted to apply the concept experimentally to more abstract 2D spaces. For example, such an abstract conceptual space would not be spanned by the physical axes “x” and “y” but for example by the conceptual axes “neck length” and “leg length” of a bird. I have to admit that I was not fully convinced by this approach, but it is interesting nevertheless.

Apart from these experimental approaches, researchers have developed theories about the representation of abstract knowledge based on hexagonal lattices (for an example theory, see Hawkins et al., 2019 or Whittington et al., 2020). Or check out this review of experimental literature that concludes that grid cells reflect the topological relationship of objects, with this relationship being defined either via space or via more abstract connections. These approaches have in common that abstract concepts are built upon the neuronal scaffold; the neuronal scaffold, in turn, is provided by the representation of physical space.

In Deutscher’s book, I found a description that paralleled the above ideas from systems neuroscience, enough to make it intriguing. First, Deutscher posits that all words that now describe abstract relationships are rooted in meanings that represent physical space. More precisely, words that were originally used to describe physical relationships (e.g., “outside”), were later taken to describe an abstract relationship (“outside of sth”, with the same meaning as “apart from sth”). We even don’t notice the metaphorical usage here because it is so common, even across languages (hormis in French, utenom in Norwegian, ausserdem in German). Deutscher highlights one specific field where the transition of spatial to more abtract descriptors is very apparant. “Before”, “after”, “around”, “at” or “on” are all prepositions that describe physical relationships but that were lateron assigned to temporal relationships as well. It seems quite obvious and not worth any further thought, but is it really?

Deutscher not only suggests that spatial relationships are more basic than temporal relationships, but he strengthens this point by pointing out that words that describe physical relationships derived from something even simpler: the own body. For example, “in front of” derives from the French word front (“forehead”). Deutscher brings up many more examples from diverse languages that reveal how the language that describes abstract relationships can be traced back to body parts. Therefore, he hypothesizes that body parts (“forehead”, “back”, “feet”, etc.) were originally used to establish a system of spatial relationships, which was then applied to temporal and other more abstract relationships. It would be interesting to investigate these lines of thought for systems neuroscience (for example, egocentric vs. allocentric coding of position, or how temporal sequences are required to define abstract knowledge representations).

One of the open questions here is about a potential interaction between abstractions, language and neuronal representations. Was this connection already implemented on the neuronal level before the inception of language, such that language only had to use this analogy generator? Researchers who work on abstract knowledge representation in animals like mice would probably say so. Or was it only through language that abstraction was enabled? I find both possibilities equally likely. Often, we cannot use a concept or idea efficiently when we don’t put it into an expression – concepts remain vague and often difficult to judge if we don’t pour them into clear thoughts, ideally written down in a consistent and concise set of sentences.

In the end, I find the ideas about the evolution of language extremely interesting, especially because they relate to the generation of abstract relationships (temporal relationships, language grammar, or fully abstract concepts). From my point of view, two aspects could be worth some further research. First, to dig into the status of the linguistic research on these topics. And second, to understand whether there are any meaningful parallels between abstraction in neuronal representations and abstraction derived within an evolving and eroding language. In any way, I can recommend this book fully to anybody who is not afraid of foreign languages and a bit of grammar.

P.S. I’ve read the German version of Deutscher’s book. It is not simply a translation of the English version but provides many additional examples from the German language that enhance or replace the examples taken from English. This was done in an excellent manner, and I can only recommend to German-speaking readers of this blog to read the translation and not the original version.

Posted in hippocampus, Neuronal activity, Review | Tagged , , | 9 Comments

Useful pieces from Twitter

Twitter used to be (and still is to some extent) a source of useful information for neuroscientists about technical details, clarifications of research findings and open discussions that cannot be obtained so easily otherwise. Here is a list of some these gems that have made it into my bookmarks, and I’m posting them here in order to archive their content at least to some detail. And yes, this list is mostly for my own reference, but it might be interesting for others as well.

Lab equipment

  1. Lior Golgher recommends to replace the expensive Conoptics Pockels cells with cheaper and better ones from Qubig, and he also makes suggestions for a fitting Pockels cell driver (A-303 or A-304).
  2. Munir Gunes Kutlu asks which camera to use to record mouse behavior compatible with posture tracking. Among the recommendations are Raspberry Pi cameras as the budget option (I have used those together with Sian Duss for pupillometry and was happy with it), the more expensive PointGrey Chameleon3 cameras; WhiteMatter cameras, where up to 15 cameras can be connected to a computer with a hub; the Basler cameras that are also used as a reliable option with few frame drops in the Helmchen lab; and the uEye XCP camera as a low-cost camera that is sufficiently good for academic purposes. It was mentioned that it is important to know beforehand whether single dropped frames are problematic or not. This is indeed important for simultaneous recording of longer chunks of behavior and neuronal activity in order to be able to synchronize two or more input streams.
  3. Tobias Rose and Jérôme Lecoq discuss where to buy cheap UV curing lights (often used for surgeries with mice to cure dental cement). Tobias bought his curing light from Aliexpress, Jérôme recommends one from McMaster-Carr and advices to use protective goggles in any case. Luke Sjulson adds that blue (not UV) curing lights are state of the art, and that they can be easily found by googling for “dental curing wand”. This made me remember a previous discussion involving Luke where curing lights and adhesive were also intensely discussed, where he recommended switching from Metabond to Optibond (both are used as the ‘the “luting” layer between bone and acrylic’).
  4. Guy Bouvier asks around for experience with iontophoresis pumps for AAV injection. I was completely unaware of this technique, which is apparently ideal for very controlled and small injections with less damage, the “gold standard for classical anterograde tracing” (according to Thanh Doan). Maximiliano Nigro seems experienced with this technique and recommends a specific precision pump.
  5. Matthijs Dorst asks about recommendations for oxygen concentrators vs. oxygen cylinders for isoflurane anesthesia stations for mouse surgeries. Oxygen concentrators seemed to work quite well, except for being rather noisy. Apart from the normal medical equipment brands (e.g., VetEquip), some reported using small fish tank pumps as a replacement for oxygen concentrators. It was mentioned that compressed air might work similarly well as compressed oxygen, but others noted that for very long procedures (>2h), mice started developing cataracts when using compressed air only. I also have learned some time ago that letting the mice inhale pure oxygen (without the isoflurane) after a surgery can improve and speed up the recovery from anesthesia.
  6. Based on some tweets (here and here), Andrey Andreev initiated a highly useful list of tips on how to set up a microscopy room (or any other lab room for a complicated data acquisition setup), including a lot of helpful links and considerations about lab infrastructure.
  7. On pre-Musk Twitter, Eleonore Duvelle used to initiate a lot of interesting discussions about rodent behavior and extracellular electrophysiology in hippocampus. She deleted these Twitter posts some time ago but is now very active on Mastodon. In addition, she has archived her Twitter threads on Github and has annotated some selected discussions in this Mastodon thread.

Software

  1. Joy A. Franco asks for tools to view z-stacks in 3D. Responses included the commercial and widely used Imaris or the free Imaris Viewer. Then, 3Dscript and ClearVolume as plugins for the open image analysis environment FIJI. Other free options included ChimeraX, AGAVE, FPBioimage. Also mentioned was Python-based napari. And, for very large (EM) datasets, there is Neuroglancer developed by Google. I wish I had the time to test all these options and compare them against each other!
  2. Maxime Beau highlights an interesting tool, the scientific Inkscape extension. Adobe Illustrator has been the best tool for vector graphic-based scientific illustrations, but Inkscape was always a good and free alternative. Affinity Designer is a newer and cheaper alternative which became more attractive when Adobe transformed to an annoying and very expensive cloud-based software a few years ago. In my opinion, all three tools are very useful, with Adobe Illustrator still being a bit better than the rest when you don’t consider the price tag. Importing PDFs generated with Python or in Matlab was, however, never easy. This extension is therefore a useful tool for any scientist using Inkscape for figure design.
  3. Matthijs Dorst asks which software/platform to use to quickly sketch out a custom imaging setup. Recommendations include gwoptics, which looks like a simple and straightforward-to-use library; more advanced, but maybe worth it for perfectionists, is Blender. A collection of example objects by Ryo Mizuta was highlighted, but there might be many more.

Other

  1. I’m no expert for sleep or sleep problems, but this thread brought up a lot of things and drugs that people apparently use to improve their sleep: amitriptyline; glycine + magnesium threonate; addition of taurine, ashwagandha, lavendar, valerian is considered; melatonin is recommended by some but not others; cannabis; antihistamines; tincture of hops; phosphatidylserine; a tidbit of morphine; trazadone; sodium axybate; gaboxadol; modafinil. I have not tried any and would not recommend anything, but I’d be curious to get all this weird stuff explained to me by an expert!
  2. Patrick Mineault asks for favorite podcasts on machine learning, neuroscience or general nerdy matters. Among the many recommendations are with a focus on neuro: Brain Inspired, the Huberman Lab Podcast, Brain Science Podcast, and the discontinued Unsupervised Thinking. With a focus on AI: Machine Learning Street Talk, TalkRL, Last Week in AI. And more broadly tuned podcasts: Sean Carrol’s Mindscape, Lex Friedman. The Twitter thread contains many more suggestions, but since I don’t like podcasts too much, I just picked some of them.

And many more threads that I forget here or that are too difficult to summarize.

Posted in Data analysis, electrophysiology, Links, machine learning, Microscopy | Tagged , , | Leave a comment

Improving the resonant scanner’s sync signal using a phase locked loop (PLL)

Calcium imaging with two-photon point scanning is the technique to chronically record from identified neurons in the living brain of animals. The central piece of two-photon point scanning microscopes is a scan engine. This can be a complex optical device like a deformable mirror or an acousto-optical deflector; but more often, it is just a mirror sitting on a rod and scanning forth and back as fast as possible. The fastest such mirrors are so-called resonant mirrors.

Currently, there is only one major provider of resonant scan mirrors for microscopy, and only few months ago, the lead times had risen to >1 year. Resonant mirrors are therefore much more precious than their mere price tag, and it is worth trying to get the best out of existing resonant scanners, instead of replacing them with new scanners at the sight of the slightest problem.

Figure 1. A typical resonant scan mirror, mounted on top of an oscillating rod. Adapted from https://novantaphotonics.com/product/crs-resonant-scanners/

Recently, I have been working with Johanna Nieweler, a PhD student in the Helmchen lab, to piece together her two-photon microscope from the remains of a previous microscope. Among the surprisingly numerous problems that Johanna encountered and fixed during this work was a problem that was ultimately due to the resonant scanner. In this blog post, I will describe how we identified the problem and came up with a – in my opinion – very elegant solution. This might be a useful resource for optical engineers who are dealing with similar problems and for microscopists who want to understand more about resonant scanners.

A periodic line jitter for high-zoom scanning

The problem was not immediately apparent. An image of small fluorescent beads acquired with the scanning microscope looked fine when zooming out:

However, to evaluate the imaging quality, but also for real experiments that resolve subcellular structures, zoomed-in imaging is essential. In our case, we noticed some kind of irregular distortion of the scan pattern, as if the beads were changing their shapes or dancing around:

To better understand this distortion, we switched off the slow galvo scanners and performed a line scan with the fast resonant scanners only. This configuration clearly revealed a periodic jitter of the scan phase of the resonant scanner.

Such an artefact could be due to many possible sources. First, it could be that the software to acquire and bin the incoming data stream might have a bug. Second, there could be a vibration responsible for sample movement. Third, it could be some line noise coupling into the “sync signal” emitted by the resonant scanner. Fourth, the mechanical scanning itself could be governed by this period modulation.

Finding the problem

The resonant scanner under scrutiny was a 4 kHz resonant scanner from Cambridge Technology. First, we measured the periodicity of the signal distortion – the frequency was around 270 Hz. So it was unlikely to be line noise, which is at 50 Hz or multiples thereof in Europe.

Next, we also did not find any vibrations that might have caused this problem.

We checked and replaced the the power supplies of the resonant scanner, without any improvement.

Finally, we turned our attention to the resonant scanner. The resonant scanner produces a so-called “sync signal”, which is a TTL signal that indicates whether the scanner is moving in the clock-wise or counter-clockwise direction. At the turning point of the directional change, the scanner flips the TTL signal and therefore generates the electrical trigger signal for the next line of the imaging frame. This means that an imprecise generation of the TTL signal, or a wobbly oscillation of the mirror itself could generate a jitter of the TTL signal and a modulation of the line signal as we observed.

Indeed, when we looked at the sync signal on the oscilloscope (we triggered on a rising flank and looked at the oscilloscope on the subsequent rising flank of the TTL), we observed a jitter of the signal that would explain the artefact in the image.

Now, this jitter could be due to a physically wobbling scan mirror; or due to an imprecise readout of the turnaround point by the TTL signal generator at the resonant scanner. Is it possible to distinguish between these two options? Yes, it is. If the scanner is scanning properly and only the TTL signal generation is affected, one could in theory simply replace the periodic TTL signal and then observe an artefact-free image.

We therefore replaced the bad TTL signal with an artificial TTL produced by a signal generator, at the exact same frequency as the resonant scanner’s frequency. Fascinatingly, we observed that this procedure fixed the problem beautifully. One can also notice that the picture drifts away if the signal generator frequency does not exactly match the frequency of the resonant scanner:

Video 1. Linescan of beads, with a signal generator replacing the sync signal of the resonant scanner. While recording the video, we slightly modified the frequency of the signal generator, resulting in a drift of the image to the left or right for even the tiniest mismatch between the scanner’s frequency and the signal generator’s frequency. We did not move the sample or the microscope at any time during the recording.

Despite this limitation, we concluded that the scanner was apparently not wobbling around, and only the generation of the TTL signal was defective.

From a workaround of the problem to a permanent and user-friendly solution

But there is a problem – we cannot simply use the signal generator with a fixed frequency TTL signal and hope that things will be fixed. First, the resonant scanning frequency of any resonant scanner changes slightly over time as it warms up. Second, the resonant scanning frequency also varies (a few fractions of a percent) when the zoom level of the microscope and therefore also the scan amplitude of the resonant scanner is changed. A mismatch between the scanner’s frequency and the sync signal would result in the drift that is visible in the video above.

What we therefore needed was a system that uses the jittery TTL sync signal of the resonant scanner and produces an output that is phase-locked to the sync signal, but without the jitter …

At this point, I vaguely remembered that this can be achieved by a simple analog electric circuit, and after an internet search, I found the Wikipedia article on the “phase locked loop” (PLL), which described exactly what we were looking for. A PLL uses an input (in this case, a jittery TTL signal) and creates a synchronized TTL signal that is typically phase-locked and runs at the same frequency but without jitter when the feedback is filtered appropriately. So we only had to implement this circuit, insert this device between the scanner’s sync signal and the “line trigger” input channel of the DAQ board, and our images would look perfectly line-triggered!

The problem here is that neither Johanna nor I had been trained as electrical engineers, so we would have struggled to translate this relatively simple idea into a working device. However, there was a true expert and lover of electrical circuits at the Brain Research Institute in Zürich: Hansjörg Kasper. Working as a support engineer at the institute for several decades, he was not only an expert for basically any practical technical question ranging from laser physics to soldering techniques, but he was also very familiar with analog circuits. While nowadays often replaced by micro-processor-based solutions, analog circuit elements such as PLLs had been standard components of electrical engineering 30-40 years ago and therefore very familiar to Hansjörg. For example, he had used PLLs in a system that he designed more than twenty years ago to track head and eye movements of small animals.

Figure 2. A photo of Hansjörg Kasper in 2011.

Hansjörg was quickly intrigued by our problem and accidentally had such a phase-locked loop circuit at his hand, the “classic” – as he called it – CD4046 PLL element.

Figure 3. Photo of a CD4046 PLL circuit, https://www.ti.com/product/CD4046B.

By the way, these circuits are quite cheap (<10 Euros). Within one day, Hansjörg soldered the components together to produce the desired behaviour. To this end, he followed the instructions that come with such a circuit element, indicating how to choose resistors and other elements to produce the desired behaviour.

He sent me this circuit diagram of his final circuit, optimized for stabilizing the sync signal of a 4 kHz resonant scanner:

Figure 4. Circuit diagram of a PLL circuit based on a CD4046 element. An inverter (“U2A 74HC14”) is included to generate a signal with the same phase shift as the input signal. At the bottom, it is indicated how to re-dimension the capacitor C1 in order to optimize the circuit for a 8 kHz instead of for a 4 kHz resonant scanner. Drawing and annotations were done by Hansjörg Kasper in KiCad.

Soldering this circuit requires a bit of practice but could be learned rather quickly by any talented tinkerer. The true challenge, in my opinion, is to start with the CD4046 datasheet (or the datasheet of a similar PLL circuit) and figure out how everything must be connected. It is not much more than 5 pages of the datasheet that are relevant, but I would probably not have understood it easily without Hansjörg’s explanations. Hopefully the circuit diagram above provided by Hansjörg will make it easy for anybody who will try to replicate our PLL stabilizer!

How it works

The circuit turned out to work beautifully. We hooked it up to the microscope, and within days we had already almost forgotten that it existed – this is a true hallmark of great engineering: that the problem is solved so perfectly and robustly that you quickly forget about its mere existence. Here is the stabilized line scan, in direct comparison with the line scan without stabilization. The residual slow wiggling of the line is due to 50 Hz line noise (a different problem).

And the same FOV with a regular scan pattern, which showed the bead without the additional dance moves, therefore enabling us to clearly see the point spread function:

If there is anybody out there struggling with a similar resonant scanner problem, I hope that this blog post will give them the tools to address and solve this problem!

In the end, I was curious whether the same solution could also be helped to generally stabilize resonant scanners. Resonant scanners are known to become “wobbly” with age or when scanning with low amplitudes (high zoom-in). I was simultaneously working also on another two-photon resonant scanning setup with an 8 kHz scanner, and I noticed some sort of wobblying and instability for very high zoom settings. Therefore, I used a signal generator to provide a highly stable sync signal to replace the scanner’s sync signal. Unfortunately, no improvements of the imaging quality could be observed. Apparently, in this case the resonant scanner was indeed wobbling physically, while the 4 kHz scanner was oscillating properly and only the generation of the TTL signal was compromised.

Altogether, this small engineering project with Johanna and Hansjörg was, in my opinion, extremely interesting and valuable. Growing up in the age of Arduinos and Raspberry Pis, where every problem can be solved by a bit of code running on a microprocessor, it was impressive to be reminded of the power of analog circuits. Of course, this implementation was only possible because we had an expert for and lover of analog circuits, Hansjörg Kasper, in our institute.

P.S. Hansjörg had been at the heart of the institute for more than 40 years. He retired officially in 2022 but did the work for this PLL project while working part-time at the Brain Research Institute in the Spring of 2023. Unexpectedly and unfortunately, he died in the early Summer of 2023.
During his time at the institute, his work had helped many dozens of PhD students and postdocs tremendously, by solving many small and big technical challenges that typical scientists are rarely equipped to address by themselves. Many a small device and machine that was built during his era, from synchronization boxes for behavioural setups to our small PLL circuit, will continue to run in the labs of the Brain Research Institute in the future, many years after him.

P.P.S. Big kudos to Johanna Nieweler, together with whom I worked on this project, to Hansjörg Kasper (R.I.P.), who designed and built the PLL circuit, to Martin Wieckhorst, who helped with the first brainstorming about the PLL circuit, and to Fritjof Helmchen, who supervises both Johanna and me.

Posted in Calcium Imaging, Imaging, Microscopy, neuroscience | Tagged , , | 6 Comments

Two random but interesting papers on neurophysiology with behavior

Bagur, Bourg et al. (2022) from the Bathellier Lab in Paris make the interesting finding that the auditory code is represented as temporal sequences of neuronal activity in early auditory processing stages and as spatial patterns in auditory cortex. I am not an expert for auditory cortex, but I found this paper quite compelling and interesting, both in terms of methodological approaches and in terms of concepts.
Sophie Bagur has already published a few very interesting papers that are both excellent and creative and definitely worth a read. For example, in a study published with her as senior author Verdier et al. (2022) on an entirely different topic, they showed that optogenetic stimulation of the medial forebrain bundle could be used to improve behavioral training of mice and replace water restriction protocols very effectively.

Lai, Tanaka et al. (2023) from Albert Lee’s lab wrote a very cool paper where they made rats navigate mentally. In short, they used the spiking output of the hippocampus, which they extracted online, to control a virtual reality that (1) represented themselves in a spatial context (“mental navigation”), or that (2) represented another object in a spatial context (“telekinesis”). Using this approach, the authors tried to show that hippocampus represents mental navigation of the self but also mental navigation of objects in the external world.
A minor issue I had with the paper was that the authors write very prominently that they constructed a low-latency (“real-time”) closed-loop system, feeding back neuronal activity into the virtual environment. However, this ultra-fast feedback is first going through a decoder that uses a 5-s running window, basically eliminating the speed of the feedback loop. This secret is hidden somewhere in the Methods section.
In addition, I had the feeling that the claim of “mental navigation” would require slightly stronger evidence that the animal did not really move physically while performing “mental navigation”. The existing analyses could have been deepened a bit. However, in general I really liked the paper, both for its ideas and for its technical implementation of the brain-machine interface for rats. Definitely worth a read!
Let’s try to dig a bit into the conceptual space around this study. The manuscript made me wonder how different it is whether we mentally move objects (like tetris objects, “telekinesis”) or avatars (like in a 1st person video game, “mental navigation”). Do both situations feel the same for us? We do certainly identify with avatars viewed from a 3rd person perspective (as in many video games). The transition from 3rd person avatars (“mental navigation”) to movable objects (“telekinesis”) seems pretty smooth from my perspective. A pixelized avatar from super mario feels almost like an object, and we treat pacman for sure like an avater, even though it is barely more than an object. There does not seem to be a sharp transition between an eating blob like pacman or snake to something less animated like a tetris element. It seems like introspection might be a good way to better understand how we as humans differentially treat virtual avatars or virtual objects in our mind.

References

Bagur, S., Bourg, J., Kempf, A., Tarpin, T., Bergaoui, K., Guo, Y., Ceballo, S., Schwenkgrub, J., Verdier, A., Puel, J.L., Bourien, J., Bathellier, B., 2022. A spatial code for temporal cues is necessary for sensory learnin. https://doi.org/10.1101/2022.12.14.520391

Lai, C., Tanaka, S., Harris, T.D., Lee, A.K., 2023. Mental navigation and telekinesis with a hippocampal map-based brain-machine interface. https://doi.org/10.1101/2023.04.07.536077

Verdier, A., Dominique, N., Groussard, D., Aldanondo, A., Bathellier, B., Bagur, S., 2022. Enhanced perceptual task performance without deprivation in mice using medial forebrain bundle stimulation. Cell Rep. Methods 2, 100355. https://doi.org/10.1016/j.crmeth.2022.100355

Posted in Brain machine interface, closed-loop, Data analysis, electrophysiology, hippocampus, machine learning, Network analysis, Neuronal activity, neuroscience, Reviews | Tagged , , , , | Leave a comment

Interesting papers on behavioral timescale synaptic plasticity (theory)

Behavioral timescale synaptic plasticity (BTSP) is a form of single-shot learning observed in hippocampal place cells in mice (Bittner et al., 2015, 2017). This finding is both interesting and inspiring for computational neuroscience for several reasons. In the first place, synaptic plasticity is the basis for learning, and any neuronal theory on the organizing principles of the brain must consider synaptic plasticity. In addition, previous findings on plasticity were usually based on artificial stimulation protocols in slices (LTP, LTD, STDP), often requiring unphysiological stimulation intensities, careful selection of stimulation protocols and many repetitions, while BTSP is a single-shot learning paradigm. Here are a some quite recent interesting theory papers on BTSP, complementing a previous blog post about experimental studies on BTSP:

Nonlinear slow-timescale mechanisms in synaptic plasticity

This review by O’Donnell (2023) nicely covers both experimental results and modeling work about the potential mechanisms of synaptic plasticity on a slow timescale. O’Donnell attempts to reconcile seeming discrepancies between experimental results on LTP from a theory perspective, and he stresses the importance of non-linearities to induce plasticity specifically during bursts but not regular spiking. A quick and interesting read!

One aspect I found particularly interesting is when he points out the relevance of long timescales and non-linear processes for classic slice LTP/LTD or STDP protocols (pages 2-3 of the PDF). Typically, one forgets about the long timescale and only focuses e.g. on the millisecond timing difference for the STDP protocol.

How BTSP affects subthreshold representations of place

Milstein et al. (2021) perform some really nice experiments and analyses to show that BTSP does not only induce new place fields but can alter existing synaptic weights in both ways, potentiation and depression.

If you read the abstract of the Milstein et al. study and have no idea what this is about: worry not, the rest of the paper is much easier to digest and actually super interesting. After reading the abstract myself, I was rather confused. One of the problems is that the “bidirectional plasticity” mentioned in the title can be misread as plasticity among bidirectionally connected cell pairs, while here the “bidirectional” only refers to “depression” (down-direction) and “potentiation” (up-direction) of synpases.

Two things I like in particular about this paper: Fig. 3L, which summarizes the main finding in a very nice visualization. And the experiments reported in Fig. 4A-K, where the authors try to disentangle the contributions of cellular depolarizations vs. initial synaptic weights for plasticity – a beautiful perturbation experiment.

P.S. This is not exactly a very “recent” paper, but I found it useful as context to better understand the subsequent paper by Moldwin et al. (2023).

A calcium level-based, general framework for plasticity

Moldwin et al. (2023) describe a framework for synaptic plasticity focused on calcium levels. It is based on the well-established finding that (a) synaptic depression occurs for low calcium levels and (b) synaptic potentiation occurs for high calcium levels, while (c) synapses remain approximately stable for the lowest (baseline) calcium levels. With their modeling work, the authors extend on existing models and make them more flexible and more easily parameter-adjustable.

In general, I think it is interesting to generate such a more general and flexible framework. At some points, however, I had the feeling that adding more model features to fit all use cases does not necessarily make the model better (more powerful yes, but not always more useful). For example, the authors write that some experimental edge cases were “incorporated into [their model] by adding additional thresholds into the step function”. Adding parameters to account for special cases does not seem the most elegant way if this additional complexity does not make any further predictions about model behavior. Despite these concerns, however, I believe that their general model based on calcium as a single actor is interesting, and their attempt to apply the model to diverse plasticity protocols is useful. It also helps to understand which assumptions are necessary to make the calcium-based plasticity model work for a particular experiment.

In the title, the authors claim that this framework could also be used to describe BTSP. Their take on BTSP is similar to the one described by Milstein et al. (2021) (see above).  Moldwin et al. (2023) assume that there is a summation of synaptic calcium induced by pre-synaptic stimulation and calcium induced by the plateau potential. The summed calcium signal would then induce LTP or LTD at this synapse. The main difference compared to Milstein et al. (2021) is therefore the idea that both slow signals, the eligibility trace and the instructive signal, are encoded in the local calcium concentration; apart from that, the mathematical shape of their model seems to be pretty similar compared to Milstein et al. (2021) as far as I can judge.

The main problem of this application to BTSP is the decay time of calcium at synaptic locations. The authors make the following assumption: “The synaptic calcium decayed with a time constant of ~2 seconds”. This seems really unrealistic – calcium levels after synaptic activation typically decay faster by an order of magnitude. Most experimentalists would assume that this long timescale is taken care of by a slow turnover of more stable mediators like CaMKII (for potentiation) or Calcineurin (for depression); see also the review by O’Donnell (2023). Overall, I have the impression that this framework is a nice and generalizable model for LTP, LTD and STPD, but maybe less so for BTSP. It is unclear to me why they advertise their model to cover BTSP so prominently in the title – this aspect may not represent the best part of their work.

A review on normative models on synaptic plasticity

In their review paper, Bredenberg and Savin (2023) discuss the criteria that should be used to judge normative models on synaptic plasticity. What do they mean with “normative models”? The authors nicely describe a very clear distinction between phenomenological and mechanistic models on the one side, which use fitting or equations to describe experimental observations; and normative models on the other side, which try to understand the importance of the modeled process for the organism. These normative models therefore ask: how should synaptic plasticity be like in order to serve the organism well? That’s an interesting perspective, not driven bottom-up by experimental observations but top-down from “desiderata”, as the title of this review indicates.

The review is interesting and establishes – not surprisingly given its focus on the practical usefulness of plasticity models – many connections to deep learning-based approaches. I was, however, slightly disappointed by the apparent lack of discussion of behavioral timescale synaptic plasticity (BTSP). For example, they describe how to choose an objective function for synaptic plasticity to improve performance and how to linearize it for small weight changes, which I found a very interesting point of view. It would be instructive to investigate BTSP and single-shot learning in such a model. In particular, it seems to me that BTSP – due to its single-shot learning behavior – cannot be a learning rule based on small and linear weight changes. It would therefore be interesting to describe BTSP in a normative model and, for example, to investigate how a loss function can be used to derive or describe BTSP.

Did I miss an interesting recent paper on this topic? Let me know!

References

Bittner, K.C., Grienberger, C., Vaidya, S.P., Milstein, A.D., Macklin, J.J., Suh, J., Tonegawa, S., Magee, J.C., 2015. Conjunctive input processing drives feature selectivity in hippocampal CA1 neurons. Nat. Neurosci. 18, 1133–1142. https://doi.org/10.1038/nn.4062

Bittner, K.C., Milstein, A.D., Grienberger, C., Romani, S., Magee, J.C., 2017. Behavioral time scale synaptic plasticity underlies CA1 place fields. Science 357, 1033–1036. https://doi.org/10.1126/science.aan3846

Bredenberg, C., Savin, C., 2023. Desiderata for normative models of synaptic plasticity. https://doi.org/10.48550/arXiv.2308.04988

Milstein, A.D., Li, Y., Bittner, K.C., Grienberger, C., Soltesz, I., Magee, J.C., Romani, S., 2021. Bidirectional synaptic plasticity rapidly modifies hippocampal representations. eLife 10, e73046. https://doi.org/10.7554/eLife.73046

Moldwin, T., Azran, L.S., Segev, I., 2023. A Generalized Framework for Calcium-Based Plasticity Describes Weight-Dependent Synaptic Changes in Behavioral Time Scale Plasticity. https://doi.org/10.1101/2023.07.13.548837

O’Donnell, C., 2023. Nonlinear slow-timescale mechanisms in synaptic plasticity. Curr. Opin. Neurobiol. 82, 102778. https://doi.org/10.1016/j.conb.2023.102778 (preprint)

Posted in Data analysis, machine learning, Network analysis, Neuronal activity, neuroscience, Reviews | Tagged , , , | 5 Comments

Interesting papers on behavioral timescale synaptic plasticity (experiments)

Behavioral timescale synaptic plasticity (BTSP) is a form of single-shot learning observed in hippocampal place cells in mice (Bittner et al., 2015, 2017). The idea is that post-synaptic activations (“eligibility traces”) are potentiated by a dendritic plateau potential/burst (“instructive signal”) that comes up to several seconds after the first signal. This result turned out to be one of the most reproducible findings of recent systems neuroscience, being replicable across labs and across stimulation protocols (via single-cell electrophysiology (Bittner et al., 2015) but also via optogenetic induction (O’Hare et al., 2022)). Many different labs are working in this interesting field, and here are a few recent papers on this topic.

A role for CaMKII in BTSP

Xiao et al. (2023) from the Magee lab report on a “critical role for CaMKII” in BTSP. “CaMKII” is known as a promotor fragment that is specific for excitatory neurons, but CaMKII is more than that – CaMKII is a calcium-dependent kinase (but see the recent paper by Tullis et al. (2023)) with important functions for plasticity. After phosphorylation, CaMKII remains in this activated state for several seconds and is therefore a candidate actor for BTSP, which acts on a timescale of several seconds as well. I learnt from this paper’s introduction that there is a CaMKII sensor based on FRET which revealed a shorter decay time for CaMKII’s phosphorylated state when it carried a specific mutation in CaMKII.

The authors take advantage of an established mouse line with this specific mutation, which resulted in deficits of memory and synaptic plasticity. Xiao et al. found, using in vivo patch clamp, that BTSP was prevented or reduced in these mutated mice (Fig. 3), suggesting an indeed critical role of CaMKII for the induction of BTSP.

As a side-finding, the authors describe that intrinsic burst propensities of CA1 neurons were quite different for mutants and wildtypes (unfortunately, they only show few example traces of such bursts), and they follow up on this result by characterizing the intrinsic properties in slices.
A question that came up and was partially addressed by the authors are the possible side-effects of chronically mutated CaMKII on the entire development and on other brain regions or also the hippocampus before the start of the experiment. However, the main finding of the manuscript (prevention of BTSP in CaMKII-mutated mice) seems to be solid despite these non-avoidable confounds. Of course, the study does not address how exactly CaMKII is involved in BTSP. The way to address this question will probably require using fluorescent reporters of CaMKII activity and very tedious dissection of molecular signaling pathways.

Monitoring of place cell induction via BTSP with calcium imaging

Grienberger and Magee (2022) conducted a simple and elegant study using calcium imaging that confirms and connects previous ideas on hippocampal plasticity and BTSP.
It had been known since a long time that hippocampal place cells are more abundant around more important locations, for example, reward locations. Here, the authors observe the transition from similar representation of all locations towards preferred representation of reward location using calcium imaging. They find evidence that this transition is induced by BTSP. To this end, they identify putative BTSP burst events in single neurons from calcium imaging data and track the activities of these neurons before and after such burst events. Moreover, they use optogenetics to show that input from entorhinal cortex (EC3) is required for the transformation of equal to reward-preferring representation of space. This finding is in line with the existing ideas about EC3 projections to provide the instructive signal that triggers bursts in apical dendrites and BTSP in CA1 neurons.

One of the most critical parts of this study is the identification of bursts to induce place fields using calcium imaging (in previous studies, this was usually done by in vivo whole-cell recordings). There is some uncertainty about the identification of these induction events (see the “CA1 place cell identification” subsection in the Methods section; check out a more recent preprint by the same lab, which uses much more complex criteria to detect BTSP events from calcium imaging; see the Methods subsection “BTSP Analysis” from  Vaidya et al. (2023)). It would be very interesting to thoroughly study – for example with simultaneous calcium imaging and single-cell electrophysiology in vivo -, which calcium events correspond to BTSP/burst events and which calcium events correspond to “regular” spiking. However, this seems technically too challenging, and the more ad-hoc methods to identify such events in this study seem fully adequate for the question addressed. P.S. If you find this paper interesting, check out the publicly available reviewer reports.

Voltage imaging of BTSP with optogenetic perturbations

Fan et al. (2023) use an impressive combination of voltage imaging and optogenetics techniques in vivo to study BTSP. Let me cite the abstract to give an idea about the methodological craziness of this study: “We combined genetically targeted voltage imaging with targeted optogenetic activation and silencing of pre- and post-synaptic neurons to study the mechanisms underlying hippocampal behavioral timescale plasticity.” Optogenetic induction of place cells in hippocampus has been done before, but not combined with voltage imaging. This together with simultaneous silencing of specific projections sounds like a completely crazy project. Several interesting findings: Excitability of pyramidal neurons (as measured with combined optogenetic activation and voltage imaging) did not increase upon BTSP. CA2/3 to CA1 inputs were potentiated upon BTSP (optogenetic activation of, although only contralateral, CA2/3 with a fiber during voltage imaging of CA1 neurons). And the activity of CA2/3 cells projecting to CA1 was required for BTSP (optogenetic silencing of projection-specific CA2/3 neurons while voltage imaging in CA1).

Overall, I would have loved to read more about these experiments and their interpretation.
Grienberger and Magee (2022) showed that (ipsilateral) EC3 input is necessary for BTSP, while Fan et al. (2023) showed that (contralateral) CA2/3 input is required. The idea is that both CA2/3 (for the eligibility trace) and EC3 input (for the instructive signal) are required, and both results are consistent with the standard theory of BTSP. It is only a bit surprising that even the contralateral CA2/3 input, as reported by Fan et al. was sufficient. In their study, the contralateral side was inhibited to enable fiber placement at CA2/3 and voltage imaging in CA1. This must be kept in mind because the authors, to access CA1 with 1P voltage imaging, also removed the central part of the external capsule; I wonder whether this removed part of the corpus callosum does not contain some of these contralateral projections.

A plasticity rule in cortex reminiscent of BTSP

Caya-Bissonnette et al. (2023) from Jean-Claude Béïque’s lab in Ottawa used slice physiology and modeling to show that a variant of BTSP exists in layer 5 pyramidal cells of cortex (S1) in mice. I have to admit that the often very long and complicated sentences in the Results section and the slightly overcrowded figures make it a tough paper to read, but it is worth it.

The authors use pre- and post-synaptic stimulation pairing paradigms as they were used extensively since the 1990s – but with the crucial difference that pre- and post-synaptic events were not separated by milliseconds but by 0.5-1.0 s. Using this approach, they found a potentiation of synapses that did not depend on the temporal order of pre/post stimulation. Therefore, they call this plasticity rule “associative”. The authors furthermore show that the endoplasmic reticulum (ER, reminiscent of O’Hare et al. (2022)) and several related calcium signaling pathways are involved in the latent “eligibility trace” that holds the memory of first stimulation before the second stimulation arrives, by extending the decay time constant of calcium (Fig. 4). The decay time is not extended by much, but maybe this slight increase is enough to induce plasticity.

Altogether, this is a very interesting study focusing on slice physiology. On one side, this is a limitation since the paper claims to have found a cortical variant of hippocampal BTSP, an effect that was primarily convincing because it had been shown in behaving animals. Therefore, one should critically inspect the choices of the plasticity protocol. For example, I noticed that the authors used a 10×20 Hz stimulation that seemed to me maybe a bit too strong and repetitive a stimulation and does not resemble typical BTSP events in vivo; moreover, during BTSP events, bursting and the associated plateau potentials were shown to be involved in the post-synaptic not the pre-synaptic cell in vivo (Bittner et al., 2015). Maybe I missed it, but I did not find this aspect discussed in the paper.

On the other side, the study is, despite this limitation, very interesting since it demonstrates how powerful slice physiology can be and how much we miss by only focusing on processes accessible in vivo. Moreover, it shows under which condition a classical STDP protocol can be transformed into a longer-timescale “BTSP” (Fig. 1F). To my knowledge, a plasticity rule with such long delays has not been convincingly demonstrated in cortex before, neither in slices nor in vivo. It may be a stretch to call the observed effect “BTSP”, drawing the analogy to the results seen in hippocampus in vivo, thereby gradually diluting the meaning of “BTSP”. However, this is just naming conventions – the science in the paper is interesting.

That’s it from my side about interesting papers on behavioral timescale synaptic plasticity.

Did I miss an interesting recent paper on this topic? Let me know in the comments!

References:

Bittner, K.C., Grienberger, C., Vaidya, S.P., Milstein, A.D., Macklin, J.J., Suh, J., Tonegawa, S., Magee, J.C., 2015. Conjunctive input processing drives feature selectivity in hippocampal CA1 neurons. Nat. Neurosci. 18, 1133–1142. https://doi.org/10.1038/nn.4062

Bittner, K.C., Milstein, A.D., Grienberger, C., Romani, S., Magee, J.C., 2017. Behavioral time scale synaptic plasticity underlies CA1 place fields. Science 357, 1033–1036. https://doi.org/10.1126/science.aan3846

Caya-Bissonnette, L., Naud, R., Béïque, J.-C., 2023. Cellular Substrate of Eligibility Traces. https://doi.org/10.1101/2023.06.29.547097

Fan, L.Z., Kim, D.K., Jennings, J.H., Tian, H., Wang, P.Y., Ramakrishnan, C., Randles, S., Sun, Y., Thadhani, E., Kim, Y.S., Quirin, S., Giocomo, L., Cohen, A.E., Deisseroth, K., 2023. All-optical physiology resolves a synaptic basis for behavioral timescale plasticity. Cell 186, 543-559.e19. https://doi.org/10.1016/j.cell.2022.12.035

Grienberger, C., Magee, J.C., 2022. Entorhinal cortex directs learning-related changes in CA1 representations. Nature 611, 554–562. https://doi.org/10.1038/s41586-022-05378-6

O’Hare, J.K., Gonzalez, K.C., Herrlinger, S.A., Hirabayashi, Y., Hewitt, V.L., Blockus, H., Szoboszlay, M., Rolotti, S.V., Geiller, T.C., Negrean, A., Chelur, V., Polleux, F., Losonczy, A., 2022. Compartment-specific tuning of dendritic feature selectivity by intracellular Ca2+ release. Science 375, eabm1670. https://doi.org/10.1126/science.abm1670

Tullis, J.E., Larsen, M.E., Rumian, N.L., Freund, R.K., Boxer, E.E., Brown, C.N., Coultrap, S.J., Schulman, H., Aoto, J., Dell’Acqua, M.L., Bayer, K.U., 2023. LTP induction by structural rather than enzymatic functions of CaMKII. Nature 621, 1–8. https://doi.org/10.1038/s41586-023-06465-y

Vaidya, S.P., Chitwood, R.A., Magee, J.C., 2023. The formation of an expanding memory representation in the hippocampus. https://doi.org/10.1101/2023.02.01.526663

Xiao, K., Li, Y., Chitwood, R.A., Magee, J.C., 2023. A critical role for CaMKII in behavioral timescale synaptic plasticity in hippocampal CA1 pyramidal neurons. https://doi.org/10.1101/2023.04.18.537377

Posted in Calcium Imaging, Data analysis, electrophysiology, Network analysis, Neuronal activity, Reviews | Tagged , , , | 4 Comments

Interesting papers on recent neuroscience methods

Find below three interesting methods papers relevant for neuroscience. All three of them are, in my opinion, worth a quick read.

Acoustic cameras to localize ultrasound vocalization of mice

Sterling et al. (2023) from the lab of Bernhard Englitz addressed the problem how to localize ultrasound vocalizations (USVs) from rodents, in particular mice. The localization of USVs and their reliable attribution to a specific mouse can be challenging, especially when mice are interacting snout-to-snout. The authors managed to perform sound localization with a precision of less than 5 mm (91% of USVs correctly assigned). How do they achieve this?

The authors combined two complementary approaches. In a first approach, they used an “acoustic camera” formed by an 8×8 array of 64 ultrasound microphones (https://sorama.eu/products/cam64). This approach provided spatially precise localization but was less reliable for high frequencies. In a second approach, they used an array of only 4 but high-quality ultrasound microphones (https://www.avisoft.com/). Figure supplement 1-1 nicely compares the noise spectra of these two microphone types. The second approach was more important to detect USV events, while the “acoustic camera” approach was essential for localization of the detected events. A nice methods paper!

Glyoxal fixation for improved antibody staining

In this study, Konno et al. (2023) from the lab of Masahiko Watanabe refined and tested an alternative protocol for immunohistochemistry (IHC), following up on a study by Richter et al. (2018). The authors replaced the standard fixative (4% paraformaldehyde, PFA) by a different solution (glyoxal) and find that this modified fixation protocol enables improved diffusion of and labeling with antibodies, including antibodies against important proteins such as synaptic adhesion molecules, receptor proteins and ion channels.

The authors mention a few caveats in the Discussion: Glyoxal fixation improved signal for certain antibodies but not for all, without an indication why this could be the case. The sample seemed to be less than hard and fixed for glyoxal than for PFA fixation. Moreover, they report that “ultrastructural images of neurons and synapses” were more strongly compromised by glyoxal fixation compared to PFA fixation. This last comment seems interesting, since Richter et al. (2018) in their original publication on glyoxal fixation stated that ultrastructure was better conserved in glyoxal than in PFA. I’m curious whether glyoxal might become the new standard for IHC during the next decade!

Near-infrared co-illumination protects fluorophores

Fluorescence occurs when fluorophore molecules transition into an excited state and, during the relaxation process into the ground state, release a photon. This process can go awry when the molecule transitions from the excited state into a dark (triplet) state or other non-desired states where the molecule is stuck (photobleaching) or which result in chemical reactions that generate oxidative stress (phototoxicity). As an experimentalist, it is essential to find the balance between (a) as much power as necessary (to be able to observe the structures of interest) and (b) as little power as possible (to prevent bleaching and toxicity).

In a surprising study, Ludvikova et al. (2023) find that near-infrared (NIR) light prevents photobleaching and phototoxicity. The authors performed imaging of standard fluorescent proteins (EGFP) in cell culture using one-photon widefield microscopy with an excitation wavelength of around 470 nm. They find that co-illumination with NIR light (approx. 900 nm) reduced bleaching induced by 470 nm. It is not fully clear from the study why NIR co-illumination prevents transitioning into triplet states, but the evidence that this indeed happens seems very convincing.

It would be interesting to see how this can be generalized to different illumination intensities and patterns, e.g., to confocal or light-sheet microscopy. And it poses the intriguing question whether two-photon microscopy of green fluorophores, which typically uses a wavelength in the NIR (around 920 nm), also takes advantage of this protective effect. For two-photon microscopy, illumination already occurs at around 900 nm, so it could be that the same photons that induce fluorescence also protect from bleaching/toxicity. It seems however challenging to test this hypothesis experimentally. As a sidenote, I’m reminded of my own experiments (reported in this blog post) where I found that long-lived dark triplet states seemed to be less relevant at 920 nm than predicted from experiments performed at 800 nm. I’m hoping that follow-up studies will further dissect the NIR co-illumination effect and study its relevance for other imaging modalities, especially confocal and light-sheet imaging!

References

Konno, K., Yamasaki, M., Miyazaki, T., Watanabe, M., 2023. Glyoxal fixation: An approach to solve immunohistochemical problem in neuroscience research. Sci. Adv. 9, eadf7084. https://doi.org/10.1126/sciadv.adf7084

Ludvikova, L., Simon, E., Deygas, M., Panier, T., Plamont, M.-A., Ollion, J., Tebo, A., Piel, M., Jullien, L., Robert, L., Le Saux, T., Espagne, A., 2023. Near-infrared co-illumination of fluorescent proteins reduces photobleaching and phototoxicity. Nat. Biotechnol. 1–5. https://doi.org/10.1038/s41587-023-01893-7

Richter, K.N., Revelo, N.H., Seitz, K.J., Helm, M.S., Sarkar, D., Saleeb, R.S., D’Este, E., Eberle, J., Wagner, E., Vogl, C., Lazaro, D.F., Richter, F., Coy-Vergara, J., Coceano, G., Boyden, E.S., Duncan, R.R., Hell, S.W., Lauterbach, M.A., Lehnart, S.E., Moser, T., Outeiro, T.F., Rehling, P., Schwappach, B., Testa, I., Zapiec, B., Rizzoli, S.O., 2018. Glyoxal as an alternative fixative to formaldehyde in immunostaining and super-resolution microscopy. EMBO J. 37, 139–159. https://doi.org/10.15252/embj.201695709

Sterling, M.L., Teunisse, R., Englitz, B., 2023. Rodent ultrasonic vocal interaction resolved with millimeter precision using hybrid beamforming. eLife 12, e86126. https://doi.org/10.7554/eLife.86126

Posted in Data analysis, Imaging, Microscopy, neuroscience, Reviews | Tagged , , | Leave a comment

Interesting papers on online motion correction for calcium imaging

In a living animal, the brain is moving up and down in the skull. This brain motion can be due to breathing, heartbeat, tongue movements, but also due to changes of posture or running. For each brain region and posture, specific brain movements are more likely to occur. For imaging of neuronal activity using two-photon microscopy through a window, the coverslip glass is usually gently pressed onto the brain to reduce motion. As a consequence, brain motion is typically smaller close to the glass surface and becomes more prominent for imaging deeper below the surface.

Why axial movement artifacts are bad

From the point of view of data analysis, there are two fundamentally different types of brain movements: Lateral movements (within the xy-imaging plane) and axial movement (in the axial z-direction). The first type can be corrected (except in the boundary region of the FOV) by shifting the image back to the “right place” within the plane. Axial movement, on the other hand, shifts the imaging focus to a different depth and cannot be corrected. From my experience, these axial motion artifacts are often underestimated and often ignored. But even if they are small (and barely visible by eye after non-rigid xy-motion correction), they might have an effect when an analysis averages across a lot of repetitions or “trials”. Especially for low imaging resolution and low signal-to-noise, it is often difficult to judge by eye whether something was due to z-motion or due to neuronal activity related to movement (or due to both). Imaging of small structures like axons or laterally-spreading dendrites is especially prone to movement artifacts, and there are many additional edge cases where brain movement might influence the analysis. Ideally, one would be able to correct for z-motion online during the recording. For example, when the brain moves up by 5 µm, the focus of the microscope would move up by the same amount, therefore imaging the same tissue despite the motion. Typically, z-motion of the brain is relatively slow, often below 20 Hz in mice. An online motion correction algorithm would therefore be required to work at a temporal responsiveness of, let’s say 40-50 Hz.

ECG-synchronized image acquisition

There have been several suggested methods to solve this problem. Paukert and Bergles (2012) developed a method to correct for only lateral (and not axial) movement, but the methods is still worth mentioning. They recorded the electro-cardiogram (ECG) and used the peak of the R-wave to reliable trigger image acquisition. Therefore, they did not really correct for heartbeat-induced motion but rather recorded such that brain motion was the same for all acquired imaging frames. Better imaging preparations (more pressure exerted onto the brain by the coverslip window) have made this approach less relevant for cortical imaging. However, for deeper imaging, e.g., with 3P microscopy, the stabilization through the coverslip is reduced. Therefore, Streich et al. (2021) used ECG-synchronized 3P scanning to improve image quality. They went a few steps further than Paukert and Bergles (2012) using the past heartbeats to predict – online with an FPGA – the next R-wave and to stop image acquisition around the R-wave window.

Correction of stereotypic axial motion with an ETL

A conceptually similar method has been developed by Chen et al. (2013). They also wanted to avoid motion artifacts induced by clearly defined motion-related events – not heartbeats in this case, but licking events. Depending on the brain region and the effort the mouse has to make, licking can cause huge motion artifacts. Chen et al. (2013) did not want to stop imaging during licking but instead wanted to correct the axial focus using a electrically tunable lens (ETL). They observed that licks consistently changed the focus of the imaging plane. They therefore used a lick sensor and fed its output to a corrective signal that moved the focus of the ETL in a calibrated manner. The idea is therefore based on the assumption that lick-induced movement is stereotyped. Therefore, one might have to recalibrate the correction algorithm for each mouse or for motion due to other parts of the body. However, it is, to my knowledge, the first attempt to really correct for axial brain motion in vivo.
It would be interesting to combine the recording of more diverse behaviors using a video camera with fast processing and with a predictive component as by Streich et al. (2021) to develop a more general brain movement predictor and corrector. For example, an algorithm could map behaviors to expected brain movements for 10 min and then perform recordings with perfectly predicted brain motion that can be anticipated and corrected by an ETL or another remote scanning device. Wouldn’t this be cool?

Using the sample surface as a reference

Another set of methods tries to image the brain and at the same time perform a scan of a reference object in the brain (a “guide star”), or the brain surface or the brain itself, in order to extract the desired corrective factor for scanning in real-time.
Laffray et al. (2011) used a reflective imaging modality to detect the position of the spinal cord (which is well known to move a lot!). In their approach, they only corrected for lateral brain motion, most likely because their microscope was not equipped with a z-scanning device. But it seems possible to generalize this method to axial movement correction as well. The limitation of this approach is that the brain surface is typically pressured by a coverslip and does not move so much – it is rather the deeper parts of the brain that move much more.

Scanning a reference object with an AOD scope

A technically more advanced method for real-time 3D movement correction was developed by Griffiths et al. (2020). In this study, the authors used a random-access AOD-based microscope, recording only small environments around single neurons. Therefore, post-hoc movement correction, also in the lateral direction, was not really an option. To address this problem, the authors developed a closed-loop system based on an FPGA to perform online motion correction. AODs (acousto-optic deflectors) are very fast scanners, so the scan speed is not an issue. The authors defined a 3D reference object that was scanned in parallel to the imaged neurons, and corrected the focus of the microscope according to movements of the reference object. This method seems to be very elegant. The only limitation that I can see is that it requires an AOD microscope – very few z-scanning techniques would be able to perform an ultra-fast z-scan as a reference (maybe TAG-lenses would be an option?). And AOD scopes have been long known to be very challenging to operate and align – although some companies seem to have managed to make the systems a bit more user-friendly during the last decade.

Scanning the brain with OCT-based imaging

Recently, I came across another paper about online movement correction, by Tucker et al., (2023). The authors used optical coherence tomography to measure brain motion and to online correct for it. In this paper, the authors present a method to correct for brain motion in x and y, not addressing the more relevant problem of z-motion (although this is discussed in depth as well). On a side-note, I noticed that corrections were applied by adding signals to the regular voltage signals used to drive the scan mirrors, reminding me of my simple solution a few years ago to add two analog signals with a summing amplifier. Anyway, a z-correction would be more difficult to implement in a standard microscope because it would require an existing z-scanning device (such as a remote z-scanner, a tunable lens, or a piezo attached to the objective). However, I found the idea compelling, and on paper the principle of OCT seems simple. An OCT is more or less an interferometer that uses infrared backscattered light to form an image. Imaging speed is high for the modern variants of OCT imaging, and the possible resolution seems to be around 1 μm. If you now have become interested in OCT, I can recommend this introduction, especially section 2.3 on “Signal formation in OCT”.

Open questions about OCT for online motion correction

Unfortunately, I have never seen or used an OCT microscope before. Therefore, I have no intuition how difficult it would be to achieve high-resolution OCT imaging. How does an OCT volumetric image of a mouse brain look like? Which structures in the brain generate the contrast? Is there any contrast? How long does it take to acquire such an image with sufficient contrast at high resolution? What is the price tag of a spectral-domain OCT system? Is it more or less difficult to build than e.g. a two-photon microscope?
I have the impression that the study by Tucker et al. (2023) only scratches the surface of what is possible, and that OCT imaging could be combined with 2P imaging without compromising excitation or detection pathways. If you are an expert in this technique and happen to meet me at the next conference, I’d be interested in knowing more!

P.S. Did I miss an interesting paper on this topic? Let me know!

[Update 2023-09-10: Also check out the comment below by Lior Golgher, who brings up additional interesting papers, including approaches to perform volumetric imaging in 3D, which enables post-hoc motion correction in the axial direction or the selction of a more motion-robust selection of an extended 3D-ROI.]

References

Chen, J.L., Pfäffli, O.A., Voigt, F.F., Margolis, D.J., Helmchen, F., 2013. Online correction of licking-induced brain motion during two-photon imaging with a tunable lens. J. Physiol. 591, 4689–4698. https://doi.org/10.1113/jphysiol.2013.259804

Griffiths, V.A., Valera, A.M., Lau, J.Y., Roš, H., Younts, T.J., Marin, B., Baragli, C., Coyle, D., Evans, G.J., Konstantinou, G., Koimtzis, T., Nadella, K.M.N.S., Punde, S.A., Kirkby, P.A., Bianco, I.H., Silver, R.A., 2020. Real-time 3D movement correction for two-photon imaging in behaving animals. Nat. Methods 1–8. https://doi.org/10.1038/s41592-020-0851-7

Laffray, S., Pagès, S., Dufour, H., Koninck, P.D., Koninck, Y.D., Côté, D., 2011. Adaptive Movement Compensation for In Vivo Imaging of Fast Cellular Dynamics within a Moving Tissue. PLOS ONE 6, e19928. https://doi.org/10.1371/journal.pone.0019928

Paukert, M., Bergles, D.E., 2012. Reduction of motion artifacts during in vivo two-photon imaging of brain through heartbeat triggered scanning. J. Physiol. 590, 2955–2963. https://doi.org/10.1113/jphysiol.2012.228114

Streich, L., Boffi, J.C., Wang, L., Alhalaseh, K., Barbieri, M., Rehm, R., Deivasigamani, S., Gross, C.T., Agarwal, A., Prevedel, R., 2021. High-resolution structural and functional deep brain imaging using adaptive optics three-photon microscopy. Nat. Methods 18, 1253–1258. https://doi.org/10.1038/s41592-021-01257-6

Tucker, S.S., Giblin, J.T., Kiliç, K., Chen, A., Tang, J., Boas, D.A., 2023. Optical coherence tomography-based design for a real-time motion corrected scanning microscope. Opt. Lett. 48, 3805–3808. https://doi.org/10.1364/OL.490087

Posted in Calcium Imaging, Data analysis, Imaging, machine learning, Microscopy, Neuronal activity, neuroscience | Tagged , , , , | 4 Comments

The idea for a journal “eLife-A”

The process of publishing and reviewing scientific results is often a strange and difficult-to-navigate landscape.

It would take too long to mention all the problems and pitfalls related to publishing. For example: overly high publication charges; long turnaround times for reviews and editorial decisions; incomprehensible reviews; incomprehensible manuscripts; delayed careers due to rejections and revision processes; lack of transparency of manuscript evaluation; lack of open access. Just to name what comes to my mind immediately.

Some of these issues have been partly addressed during the last decade by preprints (arXiv and bioRxiv), open reviews (check out my previous blog post on public peer reviews), non-legal mirrors of closed-source material (sci-hub) and experiments with new publication models (most prominently, the new eLife publication model).

Since the inception of eLife’s new publication model End of 2022, the first papers have now gone through this process, and I have the overall impression that the model is mostly well-received and definitely worth the experiment. The published papers now not only include the public peer reviews, but also an assessment by the editors, which summarizes and evaluates the claims of the paper. Manuel Schottdorf just pointed out to me that this digest can also be an outright negative assessment, for example concluding that “the evidence” for the claims is “considered inadequate” (full article plus assessment)! That’s impressive! Even though such an assessment might appear harsh from the authors’ side, such an open process makes science itself and its progress more transparent. The specific paper is outside my area of expertise, but I like the idea in general. The conclusions and claims of a paper should not be evaluated based on its association with a specific journal, but based on its assessments by expert reviewers.

In a post on Twitter, Aaron Milstein just put forward a suggestion to implement this assessment in a way that is also readable by evaluation and grant committees: by giving to each paper a grade between “A” and “F”.

For example, “eLife-A” would correspond to “Nature/Science/Cell” in terms of broad relevance, “eLife-B” to “Nature Neuroscience” or “Neuron”, “eLife-C” maybe to the current version of eLife, and so on.

Personally, I do not like the US grading scheme (from A to F), and I’m also a bit skeptical about mixing impact/relevance and methodological rigor/correctness – I really hate high-impact papers with strong claims that are not supported or only weakly supported by data, but I know that many people think differently. In any way, a one-dimensional metric (from A to F) would certainly benefit from improved interpretability, which is already better than what we have right now.

I could also imagine that as a further step, a retrospective re-evaluation of specific papers could be possible, with this secondary, retrospective evaluation reflecting the impact a paper had on its field. It is easier to tag an existing paper that already has a specific “rating” with a new additional “post-publication review rating”, rather than saying that “in retrospect, this Scientific Reports paper should have been published in Nature”. But this is just a side-note.

In an ideal scenario, such a publication model would be set up by an independent entity. eLife would actually be a good choice, because its publication model is already very close to this model. Alternatively, the EU funding agencies or the NIH could set up such a journal and oblige all projects funded by EU or NIH to publish in this journal.

One of the several positive and less obvious side-effects would be to prevent journal hopping. Journal hopping is the process of sequentially submitting a paper to several high-impact journals, in the hope of being lucky with one of them. This process not only wastes a lot of resources from both journals and voluntary reviewers but also comes at the cost of junior researchers whose careers are delayed to an unnecessary degree by this process.

I really hope that a publication scheme that gets rid of the journal tags and replaces them with grades (=subjournals) becomes a reality soon. I think it is a good idea.

Posted in Reviews | Tagged , | 2 Comments

Annual report of my intuition about the brain (2022)

How does the brain work and how can we understand it? To view this big question from a broad perspective at the end of each year, I’m reporting some of the thoughts about the brain that marked me most during the past twelve month – with the hope to advance and structure the progress in the part of my understanding of the brain that is not immediately reflected in journal publications. Enjoy the read! And check out previous year-end write-ups: 2018201920202021, 2022, 2023, 2024.

During the last years’ write-ups, I have been thinking about experiments to study self-organized plasticity in living brains. This year, I have been busy with preparations for the implementation of this line of research and hope to be working on it next year. Experimentally, I have been mostly occupied with – and more and more intrigued by – the study of hippocampal astrocytes and how they integrate calcium signals. In experiments connected to this project, we are now studying the role of neuromodulation and the locus coeruleus in more detail. And I’m glad that I can learn more about this interesting brain area by doing experiments myself. But I will discuss this in another blog post in more detail.

For this write-up, I want to discuss a small subfield of neuroscience that I only became aware of this autumn and that is closely related to ideas of self-organized neuronal networks: the experimental study of learning in cell cultures. How can the neuronal system of a cell culture be used to interact with a (virtual) world in a closed loop? In this blog post, I will discuss a few important papers to understand what can be studied with this model system of learning.

In conclusion, I think that this field is interesting in principle, as it provides a method to study plasticity (although this is not necessarily the primary aim of this kind of research). The method suffers from the problem, as most in vitro slice approaches as well, that experimental protocols for plasticity induction might be artificial and not related to processes in vivo.

I became aware if the field in October, when a high-profile paper on this topic came out and was prominently covered on social media – partly for making misleading claims and not citing prior research. I want to make it better and start chronologically with a seminal paper from 2001.

A network of cultured cortical neurons is trained to stop a stimulus

Shahaf and Marom showed in 2001 what can be described as “learning-in-a-dish” [1,2]. In these experiments, they grew a culture of cortical neurons on a 2D multi-electrode array (MEA) such that they could both stimulate these neurons and record from them. They used this system to provide some neurons with a specific stimulus. When the neuronal network exhibited a certain firing pattern briefly after stimulation, the stimulation was stopped. With this approach, the cultured network was embedded in a closed-cloop interaction.

Interestingly, the cultured network indeed started to increasingly show these spiking patterns that stopped the external stimulation. The observation is very easily summarized by figure 2 in their experimental paper [1]. After learning, the network is much more activate in a pre-defined time window (shaded area), thereby shutting down the stimulation:

This is a fascinating and also surprising observation. It seems as if the network is realizing that there is a stimulus, decides that the stimulus is annoying and therefore puts forward measures to stop and prevent the stimulus. Such a description is however highly anthropomorphic and does not really help to understand what is going on.

According to Shahaf and Marom, their observation shows in the first place that a neuronal network does not depend on a distinct implementation of reward or other forms of reinforcement. Instead, the network, following Shahaf and Marom, explores configurations until a desired state is reached (in this case: the state of stimulus removal) [1]. The authors discuss the implications a bit more in detail in [2] (check out section 7) but remain rather vague on the possible mechanisms of synaptic plasticity that might underlie such behavior.

A network of cultured cortical neurons interacts with a virtual world

Going slightly beyond the very simple closed loop by Shahaf and Marom, researchers from the lab of Steve Potter, which according to the lab’s website has “created the field of embodied cultured networks”, used the activity of the cultured neurons to drive the behavior of an “animal” in a virtual world [3]. The sensory feedback received by this animal in the virtual world is fed back to the cultured neurons. In this study, the focus is primarily on showing that it is possible to have a cultured network and a very simple virtual environment in a closed loop. This sounds particularly fascinating as the paper was published during a time when the movie Matrix had just come out.

Afterwards, in order to be able to evaluate learning and plasticity systematically, the lab moved to specific tasks based on this experimental design [4]. This study by Bakkum, Chao and Potter is conceptually close to the studies by Shahaf and Marom discussed above. The experimental design is depicted in this nice schematic (part of figure 1 in [4]) and shows a clearly identified mapping of the population activity onto a “motor output”.

Here, SBS (unpatterned shuffled background stimulation) serves as a kind of reward or neutral stimulus while PTS (patterned training stimulation) serves as an aversive stimulus that makes the network provide different output activity patterns. The (unchanging) context-control probing sequence (CPS) is used as a probe, and the network response upon the CPS pattern is regarded as the system’s motor output. Therefore, PTS or SBS act as plasticity- or feedback-stimuli, whereas CPS defines a probing time window to elicit the “motor output”.

The authors show that the networks could learn to provide mostly correct motor output when trained with such a paradigm. In addition, they quantify plasticity processes over time. Plasticity was higher during the learning procedure compared to baseline and remained higher for >1h after training. To summarize, the authors do not dig into mechanisms of cell-to-cell plasticity but rather provide a broader, systemic description of what is going on.

A difficult-to-grasp feature of this experimental study is the design of the patterned training stimulations (PTSs). The authors repeatedly applied different stimulation patterns (e.g., PTS1 stimulates electrodes 1, 4, 7 and 15, PTS2 stimulates electrodes 6, 8, 19, 20, 21 with specific stimulation intensities). Their algorithm was designed to choose from a pool of PTSs and select the PTS that resulted in plasticity towards a network state that generated the desired “motor” output. In 2022, most people (or at least myself) are so much used to deep learning and gradient descent that it is almost strange to read about such a random exploration approach.

Interestingly, the authors also investigated a comparable paradigm in a separate and purely theoretical study [5]. In this study, they replaced the cultured network with a simulated network and found very similar learning compared to the in vitro study. They found that this behavior depended on spike-time dependent plasticity, a cell-to-cell synaptic plasticity rule, and on STD, a short-term plasticity rule that helps to prevent large system-wide bursts.

From my perspective, it is interesting to see that the cell culture results could be replicated with a network that was based on STDP. A large part of the paper however instead focuses on the question how to choose and adapt stimuli that make the network learn the desired output. I would have been interested to get to know more about how STDP is shaped by the stimulation patterns and driven towards the desired output.

Learning by stimulation avoidance and autopoiesis

The group of Takashi Ikegami tried to better understand what is actually going on when neuronal cultures as in the experiments by Shahaf and Marom [1] learn to shut down an external stimulus. They study this effect using simulated neuronal networks and, as in work from the Potter lab [5], they identify spike-timing dependent plasticity (STDP) as an essential factor to mediate this learning process. As in work from the Potter lab, they also use an “embodied” application with a moving simulated robot that learns to avoid walls [6].

The main take-away of their simulation experiments is that the experiments by Shahaf and Marom can be replicated in silico and that the observed phenomena can be interpreted as “Learning by Stimulation Avoidance (LSA)”. The authors write in the discussion:

LSA is not just an another theoretical neural learning rule, but provides a new interpretation to the learning behavior of neural cells in vitro. (…) LSA relies on the mechanism of STDP, and we demonstrated that the conditions to obtain LSA are: (1) Causal coupling between neural network’s behaviour and environmental stimulation; (2) Burst suppression. We obtain burst suppression by increasing the input noise in the model or by using STP.

I have the impression that this is an overstatement of what “LSA” is. First, both STDP and burst suppression by spontaneous activity as ingredients to enable learning-in-a-dish had been described already previously [5]. Second, I don’t think this is “a new interpretation” of in vitro learning but simply the demonstration of a model that is consistent with the observed experiments from [1].

The authors expand a bit and provide more context:

As we have shown, LSA does not support the theory of the Stimulus Regulation Principle. It could be closer to the Principle of Free Energy Minimisation introduced by Friston. The Free Energy Principle states that networks strive to avoid surprising inputs by learning to predict external stimulation.

The “stimulus regulation principle” suggests a network that samples possible network configurations and stops the exploration once a favorable state is reached. A favorable configuration would be one which manages to suppress the stimulation. The STDP-model put forward by [6] instead is based on a synaptic learning rule that specifically strengthens and weakens synapses in order to reach a favorable network configuration.

The mentioned “free energy principle”, on the other hand, is infamous for being so general that almost any imaginable theory is consistent with it. The most popular theory about an implementation of the free energy principle is probably predictive processing [7]. In classical predictive processing, an internal model in the brain provides top-down signals to cancel out expected sensory inputs. The internal model adapts in order to improve the predictive power of the top-down signals. It is interesting how this typically hierarchical top-down/bottom-up view breaks down when being applied it to the model system of a dissociated cell culture. It would be hard to assume that there is an hierarchy in this dish. And still, the stimulus-avoidance paradigm does have some clear resemblance with surprise-minimization and predictive processing.

For more work from the Ikegami lab in collaboration with others, check out [8], where they interpret the findings that neurons in cultured networks that receive incontrollable stimulation will be disconnected from the network. They do so in the context of the rather old concept of autopoiesis, thereby going beyond the simple principle of “learning by stimulus avoidance” (LSA).

A network of cultured cortical neurons learns to play a simplified version of pong

In October 2022, a related paper by Kagan et al. with Karl Friston as senior author – who is famous for inventing the free energy princple – was published [9]. It received a lot of attention, but also quite some criticism by several neuroscientists on Twitter (ref 1, ref 2, ref 3, and other discussions that were deleted afterwards). It was critizised that the paper failed to cite previous research and interpreted the results in a misleading manner. For example, the title speaks of cultured neurons that “exhibit sentience”. This statement does not use a commonly used definition of “sentience” and therefore seems quite misleading for the reader. Unfortunately, this is not the only part of the manuscript which sounds like a buzzword combination made up by ChatGPT. Check out this excerpt of the introduction:

Instantiating SBIs [synthetic biological intelligences] could herald a paradigm shift of research into biological intelligence, including pseudo-cognitive responses as part of drug screening, bridging the divide between single-cell and population-coding approaches to understanding neurobiology, exploring how BNNs compute to inform machine-learning approaches, and potentially giving rise to silico-biological computational platforms that surpass the performance of existing purely silicon hardware.

But let’s put all the the buzz aside and discuss the science. Similar to the previously discussed studies, Kagan et al. used cultured neurons that were integrated into a closed loop using multi-electrode arrays. The virtual environment of this closed loop was a simplified version of the game “pong”. The multi-electrode array (MEA) was used to provide the cell culture with “sensory” feedback about the position of the ball and the success of the game, and to provide “motor” output from the MEA activity pattern back to the game in order to control the paddle.

Pong as implemented in the paper was less demanding (using a paddle half as large as the side of the arena). Despite this simplification, the performance of the cell culture controlling the game was not very impressive. The cultured network, after some learning, managed to hit the ball on average a bit more often than once (1x) before losing the game. This performance is not much higher (but significantly above) chance level.

The experimental design is well illustrated by this panel from their figure 5 (taken from the preprint, which had been published under the CC-BY-NC-ND-4.0 license):

The electrode array is subdivided into a “sensory” region indicated by the crosses and a “motor” region indicated by up and down arrows; the firing patterns in the motor region move the paddle up or down. The sensory region is provided with sensory feedback about the position of the ball, but also with a predictable stimulus when the ball hits the paddle and a random stimulus when the paddle misses the ball. Over time, the neurons learn to avoid the random stimulation by hitting the ball more often.

The interesting aspect is that the network seems to learn based on an intrinsic preference for predictable as opposed to unpredictable stimuli. The authors interpret this result as supportive evidence for the free energy principle, because the systems seems to learn the game in order to escape punishment by the random and unpredictable stimulus.

I have however some doubts about the interpretation of the results and also about the conceptualization. First, it is strange that predictable and unpredictable stimuli are used to reward or punish the system. This is not how, according to my understanding, the free energy principle works. One would rather expect the the system (the ball/paddle interaction) to be modeled by the neuronal network and therefore become predictable. There would not be any use of additional reinforcement by predictable stimuli as reward. – Interestingly, in Bakkum et al. [4], in contrast, the negatively reinforcing stimulus was patterned while the stabilizing, positively reinforcing stimulus was random. This fact shows that different stimuli are used for different purposes, and interpretations of their effect on the network depends on the conceptual framework of the paper.

Second, it is strange that the predictable stimulus (75 mV at 100 Hz over 100 ms; unfortunately, the duty cycle is not stated) and the unpredictable stimuli (150 mV at 5 Hz over 4000 ms) differ quite a bit in terms of stimulation strength, duration and frequency of stimulation. One needs to make oneself aware of the fact that in reality the “predictive stimulus” is a high-frequency tetanus stimulation. Such tetanic stimuli are known to be able to drive long-term potentiation (LTP). It is not hard to imagine that the slower unpredictable stimulation results in long-term depression (LTD), not by means of being unpredictable but by means of its frequency and duration.

Therefore, an alternative and, in my opinion, much more parsimonious explanation of the results is that the system does not try to minimize surprising/unpredictable results, but that it potentiates neuronal patterns that precede successful performance by tetanic stimulation. I am therefore not convinced that Kagan et al. demonstrated a phenomenon that can be linked to predictive processing or the free energy principle in the way they describe it.

However, I would also like to highlight the positive sides of the paper: the figures are well-designed; a preprint is available; the experiments were done with several conditions which allow for comparisons across conditions (e.g., different cell lines); and the experimental system was characterized with several complementary methods (including IHC, EM, qPCR).

Conclusion

The fact that cultured neuronal networks can interact with a closed-loop and learn to optimize these interactions is quite fascinating. The interpretation of such behavior as “learned stimulation avoidance” [6], “autopoiesis” [8] or a reflection of the “free energy principle” [6,9] is certainly intriguing. However, it might rather serve as a metaphor or a high-level description and does not really provide a deeper analysis or understanding of the underlying mechanisms. One possible starting point to investigate the mechanisms of such learning could be the STDP rules that were found to be consistent with the learning behavior in simulation studies [5,6]. It would be possible to make predictions about how spike sequences evolve over learning in the dish, and to test those predictions in the experiment.

It is remarkable how limited the performance of the cultured in vitro systems is when they were trained to control a virtual agent or a game. The performances are, seemingly without exceptions, barely above chance level, and nowhere close to mastering the virtual world. Years ago, deep Q-learning has achieved performances that are close to perfection in video games way more complex than “pong”. I do not think that anybody should make a point of “intelligence in a dish” when the dish can barely process a binary input.

However, I am still somehow intrigued by the experimental findings, especially those made initially by Shahaf and Marom [1,2]. Would it be possible to observe the same behavior with a more naturalistic network, for example a cultured slice, or an ex vivo explant of a brain, or even a brain region in a living animal? For example, one could use optogenetics to stimulate a subset of cortical neurons in the mouse (“sensory” neurons), use calcium imaging or electrophysiology to record from another subset of neurons in the same brain area (“motor” neurons) and stop the stimulation once the “motor” neurons show a specific activity pattern. Using this experimental design, one could test for example whether the learning by stimulation avoidance can also be elicited in vivo.

Another aspect that might deserve attention is the apparent dependence of plasticity processes on a strong stimulus. In the experiments by Shahaf et al. [1], the stimulus co-occurs with the activity pattern that is then modified by plasticity. In Bakkum et al. [4], the plasticity-inducing stimulus is a tetanic stimulation following the activity pattern. Very similarly, in Kagan et al. [9], the tetanic stimulus is directly following the activity pattern that is strengthened afterwards. The observed effect could therefore be the reflection of a very coarse form of plasticity which occurs in a network that is strongly stimulated by an external stimulus that somehow reinforces previous activity patterns.

Overall, from all these studies it becomes clear that networks of cultured neurons do indeed seem to show some level of systems-level learning when integrated into a closed loop. I have the impression that the plasticity induced (e.g., by a strong tetanic stimulation) is not very naturalistic. Despite this limitation, it would be interesting to investigate experimentally what plasticity rules underlie such behavior. It is likely that the plasticity studied in slices since the ’90s is in some similar way artificial since it does not integrate neuromodulation or realistic levels of external calcium. And still these slice experiments brought forward useful concepts that can be tested in living organisms. Similarly, I think there might be interesting aspects of studying plasticity rules in cultered neurons in closed loops. But I do not think that it is of any use to frame such cultured networks as “synthetic biological intelligence” [9] or to overinterpret the results in a theoretical framework of choice – in particular if the performance of this “intelligence” remains so much lower than the lowest standards of both biological and artificial neural networks.

References

[1] Shahaf, Goded, and Shimon Marom. Learning in networks of cortical neurons. Journal of Neuroscience 21.22 (2001): 8782-8788.

[2] Marom, Shimon, and Goded Shahaf. Development, learning and memory in large random networks of cortical neurons: lessons beyond anatomy. Quarterly reviews of biophysics 35.1 (2002): 63-87.

[3] DeMarse, Thomas B., et al. The neurally controlled animat: biological brains acting with simulated bodies. Autonomous robots 11.3 (2001): 305-310.

[4] Bakkum, Douglas J., Zenas C. Chao, and Steve M. Potter. Spatio-temporal electrical stimuli shape behavior of an embodied cortical network in a goal-directed learning task. Journal of neural engineering 5.3 (2008): 310.

[5] Chao, Zenas C., Douglas J. Bakkum, and Steve M. Potter. Shaping embodied neural networks for adaptive goal-directed behavior. PLoS computational biology 4.3 (2008): e1000042.

[6] Sinapayen, Lana, Atsushi Masumori, and Takashi Ikegami. Learning by stimulation avoidance: A principle to control spiking neural networks dynamics. PloS one 12.2 (2017): e0170388.

[7] Keller, Georg B., and Thomas D. Mrsic-Flogel. Predictive processing: a canonical cortical computation. Neuron 100.2 (2018): 424-435.

[8] Masumori, Atsushi, et al. Neural autopoiesis: Organizing self-boundaries by stimulus avoidance in biological and artificial neural networks. Artificial Life 26.1 (2020): 130-151.

[9] Kagan, Brett J., et al. In vitro neurons learn and exhibit sentience when embodied in a simulated game-world. Neuron 110.23 (2022): 3952-3969.

Posted in electrophysiology, Network analysis, Neuronal activity, Review, Reviews | Tagged , , | 11 Comments