U.S. flag

An official website of the United States government

NCBI Bookshelf. A service of the National Library of Medicine, National Institutes of Health.

Noebels JL, Avoli M, Rogawski MA, et al., editors. Jasper's Basic Mechanisms of the Epilepsies. 5th edition. New York: Oxford University Press; 2024. doi: 10.1093/med/9780197549469.003.0060

Cover of Jasper's Basic Mechanisms of the Epilepsies

Jasper's Basic Mechanisms of the Epilepsies. 5th edition.

Show details

Chapter 60Artificial Intelligence–Guided Behavioral Phenotyping in Epilepsy

and .

Abstract

A major impediment to progress in basic epilepsy research is the fact that evidence-based, rigorous translational research is not only prohibitively time and labor-intensive, such as 24/7 video-electroencephalogram recordings, but rests on inherently subjective scoring by human observers, as exemplified by the Racine scale. Recent technical progress in machine learning and computer vision highlighted a variety of novel possibilities for quantifying behavior in animal models of epilepsies. This chapter briefly reviews the latest advances in artificial intelligence (AI)-guided animal motion tracking and segmentation of pose dynamics that bear great potential of revolutionizing behavioral phenotyping in basic epilepsy research. As an emerging field fueled by the recent successes of deep learning, AI-guided behavioral phenotyping will be discussed primarily in order to provide insights into the fundamentals of the field and at the same time raise awareness of potential pitfalls of the underlying technology. By concisely surveying the diverse and rapidly growing landscape of the relevant methods and toolboxes available in neuroscience research, this chapter aims to spark interest in AI-aided behavioral phenotyping in the epilepsy community.

Introduction

Why Do We Care about Behavior in Epilepsy?

For both research and diagnosis of epilepsy, studying the manifestation of the disease most often starts by assessing the signs and symptoms that are expressed in behavior. As such, behavior is a functional readout of brain activity, and in epilepsy, it is specifically shaped by ictal and interictal activity. The spectrum of behavior associated with epilepsies ranges from uncontrolled movements (e.g., jerks or twitches), emotional (e.g., anxiety) and cognitive symptoms (e.g., memory impairments), to subtle changes in responsiveness to the environment (e.g., temporary confusion or loss of awareness). Analyzing behavior during (ictal) and between (interictal) seizures is central for classifying seizures (Fisher et al., 2017) and epilepsies (Scheffer et al., 2017), as well as uncovering comorbidities (e.g., autism [Abrahams et al., 2011; Tsai et al., 2012] or memory impairments [Bui et al., 2018]), assessing epilepsy progression and treatment efficacy in both clinical and lab settings. Together with its role (e.g., as a functional readout) in helping to uncover the causes of the epilepsies, the need for evaluating behavior can be found across the shared research priorities of the epilepsy community (Binder et al., 2020; Chang et al., 2020; Jones et al., 2020; Poduri and Whittemore, 2020; Traynelis et al., 2020).

What Kind of Behavioral Readout and Expertise Is Required to Phenotype Animal Models of Epilepsies?

A behavioral readout can serve different objectives. As a task-specific metric, behavior is used to assess known impairments by quantifying predefined labels. For example, a gold standard to evaluate memory deficit in a mouse model of temporal lobe epilepsy (TLE) is to determine and compare the exploration time (e.g., time spent at predefined distance to object) between moved and unmoved objects in the so-called object location memory task (Vogel-Ciernia and Wood, 2014; Bui et al., 2018; Kim et al., 2020). In contrast to such task-defined measurements, renouncing any a priori assumptions about relevant behavioral features, an observation-based deconstruction of naturally occurring behavior (e.g., observation of pathognomonic head movements) can give rise to a new perspective by suggesting testable hypotheses about the potential brain networks involved (e.g., the proposed [Noebels et al., 1990] and later confirmed [Khan et al., 2004] vestibular damage in stargazer mice). Historically, these seemingly dichotomous strategies to evaluate behavior, that is, assessing task-related response behavior and exposing the structure of behavior, arose from two distinct traditions, comparative psychology and ethology (Gomez-Marin and Ghazanfar, 2019). However, recent initiatives in systems neuroscience argue that these disciplines have to come together to disentangle the relationship between brain and behavior (Datta et al., 2019; Dennis et al., 2021).

For preclinical epilepsy research as well, the necessity of coalescing different methodologies to capture unique yet comparable behavioral phenotypes in animal models of epilepsies through an unbiased and data-driven manner becomes increasingly evident with more tools automating behavioral measurements at hand (Fig. 60–1). While recent technological advances make it more accessible to continuously track behavior at progressively higher resolution (Mathis and Mathis, 2020; Mathis et al., 2020; Pereira, Shaevitz, and Murthy, 2020), traditional approaches persist and still dictate how behavioral readouts are compiled and behavioral phenotypes are tessellated to fit their methodological framework. For example, while continuous video-electroencephalogram (EEG) can capture seizure burden across various timescales (e.g., ultradian and circadian dynamics; Quigg et al., 1998; Kim et al., 2020) and disease stages (e.g., epileptogenesis; Williams et al., 2009), the analysis of behavior is usually at best restricted to the ictal periods, timestamped by the EEG, while interictal periods in these datasets are largely ignored (Fig. 60–1A). Instead, in order to assess any additional, interictal impairments and characterize comorbidities, behavioral assays are separately chosen and deployed at selected time points, usually without any neural recordings (e.g., EEG). We argue that an integrative system is required that takes advantage of the recent technological advances in automated machine vision to capture the rich behavioral repertoire in animal models of epilepsy (Fig. 60–1B), combining multimodal measurements (such as video and EEG) to a holistic, functional readout capable of testing for known deficits and screening for novel impairments and drugs without observer bias.

Figure 60–1.. The potential of artificial intelligence (AI)-guided phenotyping in basic epilepsy research.

Figure 60–1.

The potential of artificial intelligence (AI)-guided phenotyping in basic epilepsy research. A. In classical approaches, behavioral phenotypes in animal models of epilepsies rely on separate measures of behavior, often segregating the acquisition of (more...)

As with data collection in animal models of epilepsy, the availability of new analysis tools, in particular for animal behavior and tracking, has rapidly grown and is likely to continue to do so over the next decades, thanks to the tremendous success of artificial intelligence (AI). AI has a great potential to facilitate and accelerate several aspects of basic epilepsy research, especially, as discussed below, for behavioral tracking and analysis. However, the shift from manual analysis of small datasets acquired by an individual research lab toward an AI-automated, large-scale computation of “big-data” with shared data structures (Teeters et al., 2015; Rübel et al., 2019) and datasets (Abbott et al., 2017; Bonacchi et al., 2019) requires the acquisition of new skill sets by the research community. With a growing landscape of tools with variable degrees of manual human contribution and intelligible insight into the algorithms used, it remains the burden of the end user to acquire a basic knowledge of the underlying technology, if only to gain awareness of its pitfalls. However, we argue, due to the growing presence of AI and a large body of literature about animal tracking, a basic knowledge about AI (including terms and functionalities) and an overview of current state-of-the-art behavioral tracking approaches in neuroscience will help both researchers and future tool developers to either use or adapt the technology for basic epilepsy research. Therefore, in this chapter, we will discuss the idiosyncrasies of AI with a focus on behavioral tracking approaches that bear potential to significantly benefit basic epilepsy research. Since other chapters in this book discuss more traditional behavioral studies in various animal models of epilepsies, here we will focus on pre–machine learning and machine learning–guided approaches for animal tracking and behavioral analysis in laboratory settings, while highlighting their potential for AI-guided behavioral phenotyping in epilepsy.

Analyzing Behavior Starts with Tracking

In epilepsy research and for neuroscience in general, studying the functions and dysfunctions of the brain, many have argued, requires a careful decomposition of behavior (Krakauer et al., 2017; Datta et al., 2019; Dennis et al., 2021). Describing an animal’s behavior in its environment usually starts by tracking individual body parts and analyzing the pose dynamics through time. Hence, it is important to remember that the features that are tracked directly define the granularity of behavioral description (Fig. 60–2). A large body of research in computer vision and more recently in neuroscience has focused on accurately and efficiently tracking key points up to entire poses of animals in videos (e.g., reviewed in Mathis et al., 2020; J. Wang et al., 2021). In parallel, another branch of research has focused on developing better tools to quantitatively assess behavior by segmenting extracted features (e.g., location and acceleration of the animal’s paws) into stereotypic, recurring units (e.g., grooming; reviewed in Datta et al., 2019; Pereira, Shaevitz, and Murthy, 2020). Both fields have benefited from several recent advances in machine learning (ML) and have started to tackle more complex scenarios (e.g., social interactions using multianimal tracking; Pereira et al., 2020; Lauer et al., 2021). Together with recent efforts to integrate multimodal measurements (e.g., ultrasonic vocalizations [USVs]; Karigo et al., 2021), progress in different disciplines deploying ML techniques (including computer vision or speech recognition) will help capture an animal’s interaction with its environment in greater detail and ultimately help to understand how the brain generates behavior or fails to do so. In the following paragraphs, we briefly discuss the origins of animal tracking algorithms and their advances toward estimating poses, methods used to segment pose dynamics into stereotypical behavioral modules, and other, less-known approaches with great potential that try to capture and model entire scenes.

Figure 60–2.. Animal tracking—from point tracking to 3D surface reconstruction.

Figure 60–2.

Animal tracking—from point tracking to 3D surface reconstruction. A. Behavior involves coordinated movements; thus, the features that are tracked defined the granularity of behavioral description. Experimenters need to decide what tracking mode (more...)

Tracking animals has a long history and is used in several settings (e.g., neuroscience research, animal husbandry, industrial farming, or animal conservation). Besides being driven by a particular question—scientific or other—choosing a tracking approach often depends on the circumstances (i.e., on the particularities of the animal and the environment), in which the data has to be acquired (e.g., livestock tracking on a farm or tracking movements of an octopus in an aquarium). Thus, different approaches can be broadly categorized by the type of sensors that are used, such as videography, microphone, radio-frequency identification (RFID) tags, and inertial measurement unit (IMU) sensors. For neuroscience research, which largely takes place in controlled laboratory settings, video-based approaches have been the most fruitful for studying the behavior in model organisms and have been accompanied by several technical breakthroughs over the last two decades in particular regarding invasiveness (e.g., from marker-based to marker-less tracking) and richness of the readout extracted (e.g., from 2D position to 3D pose tracking).

Early computer-assisted approaches to track animals in videos focused on simple readouts like position and speed, which only required a single point to represent the animal (e.g., the animal’s centroid) (Fig. 60–2B). Such approaches often rely on background subtraction, removing pixels in an image belonging to the environment in order to identify the animal. In laboratory settings with static environments, one of the most common and fastest approaches, and thus also suitable for real-time applications, is to take advantage of contrast or difference in color hues (i.e., the chroma range) between animal and background by building behavioral arenas with a light background (Voigts, Sakmann, and Celike, 2008) or including a green screen (Maghsoudi et al., 2018; Haji Maghsoudi, Vahedipour, and Spence, 2019; Bala et al., 2020), while applying image intensity thresholding or chroma keying, respectively, to separate foreground (i.e., the animal) from the background. While such approaches do not constrain an animal’s behavior per se, they impose experimental settings with limited natural context and are prone to fail in dynamic scenes (e.g., with multiple animals) without the use of any additional markers or sensors (e.g., RFID tags).

With a growing research interest in quantifying an animal’s interaction with the environment (e.g., exploration of objects [Vogel-Ciernia and Wood, 2014] or hunting prey [Johnson et al., 2020]) and thus reasoning about an animal’s sensory perception about what lies ahead (e.g., through smell and vision), tracking the orientation of an animal became more important, especially since virtually all model organisms in neuroscience (from Caenorhabditis elegans to Mus musculus) share a bilaterally symmetric body plan with an anterior-posterior (head-to-tail) and a ventral-dorsal (belly-to-back) axis. From a technical point of view, however, distinguishing individual body parts to decipher the orientation and eventually the pose of an animal brought several new challenges of which some are solved and others remain to this day. Tracking body parts reliably despite obstruction (e.g., by objects in the scene or during grooming) or the lack of unique features (e.g., head vs. tail in C. elegans) was traditionally addressed by combining classical video-based approaches with markers (e.g., coloring body parts [Spink et al., 2001; Inayat et al., 2020] or affixing retro-reflective markers [Roy et al., 2011; Mimica et al., 2018; Marshall et al., 2021]) or special sensors (e.g., RFID [Peleh et al., 2019] or IMU [Pasquet et al., 2016; Vanzella et al., 2019]). While such marker-based approaches are highly accurate (e.g., as shown for tracking whisker and measuring head rotations) and usually do not need any further postprocessing of the data (e.g., manual annotation), they are widely deemed impractical and at worst interfere with the natural behavior of an animal. More recently, like other classical computer vision approaches (e.g., optical flow [Horn and Schunck, 1981; Hur and Roth, 2020]), marker-based methods are mostly replaced with a deep learning (DL) approach. Nevertheless, marker-based methods still serve an important role as valuable ground truth data, providing labels for training ML–based methods (see Dunn et al., 2021, for an excellent example of such a transition from marker-based to marker-less, ML–based motion capture in rodents).

Machine Learning and Deep Learning Revolutionized Animal Motion Tracking

With Machine Learning Toward Marker-less Animal Tracking

AI as an academic discipline is rather young (1950s), but nonetheless it has shown its transformative potential in several areas, including basic neuroscience research, where it reduces manual labor for behavioral analysis in animals, while increasing reliability and accessibility of accurate pose estimation in different experimental settings (e.g., see Mathis et al., 2020; Pereira, Shaevitz, and Murthy, 2020). AI comprises different disciplines, including ML, the study of algorithms designed to improve over time by extracting patterns from raw data (Goodfellow, Bengio, and Courville, 2016). Computer vision, like other scientific fields in and around computer science (e.g., natural language processing), adopted ML approaches early and used hand-designed features in combination with a (machine) learning subsystem (e.g., classifiers) for identifying and classifying image features. For example, image features, such as histogram of gradients (HOG), which describes an image based on the distribution of pixel intensity gradients and edge directions, are extracted and vectorized into a feature vector to train ML classifiers like a support vector machine (SVM) in order to classify images into different categories (e.g., photos with and without humans) (Dalal and Triggs, 2005). Although such classical ML pipelines had great success in tasks like image classification and gave rise to great tools for animal tracking that simplified, for example, orientation tracking (Fig. 60–2C) (Branson et al., 2009; Dankert et al., 2009), these classical techniques required a fair amount of feature engineering (e.g., HOG) and domain knowledge (e.g., in image preprocessing such as image filtering) to design suitable representations (i.e., feature vectors describing the raw data in compressed form) for the downstream ML algorithm. Other preprocessing algorithms in computer vision such as simple linear iterative clustering (SLIC) (Achanta et al., 2010), which clusters pixels based on their color similarity and proximity in the image into superpixels (Ren and Malik, 2003), tried to reduce redundancies in the raw image itself instead of extracting specific hand-crafted features. Combined with an algorithm like conditional random fields (CRFs), SLIC had great success in tasks like image segmentation, which is useful for background subtraction or (body) part segmentation (i.e., semantic segmentation). Although image preprocessing algorithms such as SLIC and others are used for animal tracking in laboratory settings (Kyme et al., 2014; Machado et al., 2015; Maghsoudi et al., 2018; Haji Maghsoudi, Vahedipour, and Spence, 2019), nowadays their main utility lies more in their processing speed (e.g., for real-time applications) rather than providing the best features for a downstream ML algorithm.

Basics of Deep Learning

The quest for good feature extractors that transform raw data into suitable representations largely ended with the emergence of efficient artificial neural networks (ANNs) that are able to learn directly from the raw data what representation is needed, for example, for detection or classification (Fig. 60–3). While DL comprises a large family of ML approaches that are based on ANNs with multiple layers (hence the term “deep”), DL architectures such as deep neural networks (DNNs) became particularly popular in tasks such as image classification and segmentation. Structurally, DL methods such as DNNs are composed of multiple processing layers with simple, nonlinear units that transform the raw input data into multiple representations with gradually higher levels of abstraction (LeCun, Bengio, and Hinton, 2015; Goodfellow, Bengio, and Courville, 2016). One of the main reasons for the great success of DL approaches came with their capability of being trained “end to end” by using a single multilayer neural network to learn the mapping of raw input data (e.g., array of pixels in an image) to an output (e.g., scores for categories like “cat” or “dog”). Although it is not the objective to provide a comprehensive introduction to DL here, in the following paragraph we will introduce a few general concepts and useful terms that hopefully will help researchers without DL experience navigate through the growing body of literature in animal tracking.

Figure 60–3.. Deep learning basics.

Figure 60–3.

Deep learning basics. A. An example of a multilayer neural network with an input, two hidden, and an output layer. Such feedforward neural network can be trained with stochastic gradient descent in three main steps: forward pass, loss computation, and (more...)

While there are many different categories of DL algorithms, those for supervised learning are probably the most common. Supervised learning algorithms require training with large datasets that consist of several thousand or more of such input-output pairs with the objective (represented by an “objective function”) to reduce the error between predicted and desired output by iteratively adjusting internal parameters, the so-called weights. In practice, training a feedforward neural network in a procedure called stochastic gradient descent (SGD) can be divided into three main steps (Fig. 60–3A): forward pass, loss computation, backward pass. First, the output scores (“prediction”) are computed by (forward) passing the raw data from the input layer through the hidden layers to the output layer, computing for each unit in a layer the total input from the units of the layer below as a weighted sum (“weights”) and then passing the result through a nonlinear activation function (such as a rectified linear unit [ReLU]). Second, the error (or loss) between predicted value and ground truth is calculated by an objective function (also known as loss and cost function), which “can be seen as a kind of hilly landscape in the high-dimensional space of weight values” (LeCun, Bengio, and Hinton, 2015). Finally, the negative gradient, “the direction of steepest descent in this landscape” (LeCun, Bengio, and Hinton, 2015), is calculated by backpropagation, which can be used to adjust the weights and is merely a practical application of the chain rule for derivatives from calculus, a technique to find the derivative of composite functions. While the use of backpropagation to compute gradients was discovered in the 1970s and 1980s, neural networks were unpopular in the 1990s until the early 2000s, when they experienced a renaissance with deep feedforward networks and the emergence of fast graphics processing units (GPUs) (e.g., see review by LeCun, Bengio, and Hinton, 2015, or Bottou, Curtis, and Nocedal, 2018).

A subtype of deep feedforward networks that became particularly successful during the early 2010s were convolutional neural networks (ConvNets [CNNs]), especially in tasks requiring image recognition, segmentation, and detection (Krizhevsky, Sutskever, and Hinton, 2012; Farabet et al., 2013; Girshick et al., 2014; Sermanet et al., 2014). Key properties of ConvNets include local connections to extract highly correlated features, shared weights that can detect location-invariant features, and pooling to merge semantically similar features together (Fig. 60–3B). With the alternation of convolutional and pooling layers, ConvNets are capable of decomposing image features hierarchically, and thus, resembling the computation in the ventral visual stream of the mammalian brain (Hubel and Wiesel, 1962; Felleman and Van Essen, 1991). Interestingly, in computational neuroscience, models based on ConvNets were even shown to predict single-unit and population responses in higher visual areas (e.g., in V4 and inferior temporal cortex) to naturalistic images and thus ConvNet models turned out to be useful for studying high-level human abilities such as visual object recognition (Yamins et al., 2014). More generally, ConvNets can be used to study sensory cortical processing, modeling the encoding process that transforms external stimuli into neuronal activity; in contrast to the decoding process where neuronal activity generates behavior (Yamins and DiCarlo, 2016). From a practical perspective, an encoding-decoding architecture with ConvNets (Fig. 60–3B) is particularly useful in combination with transfer learning (Caruana, 1994; Bengio, 2011; Bengio et al., 2011; Yosinski et al., 2014). With architectures such as DeepLab (Chen et al., 2018), for example, this allows a neural network to be extensively pretrained on a large dataset such as ImageNet (Deng et al., 2009), and then the trained encoder (“backbone”) can be transferred to a custom network, where only higher-level portions (e.g., the decoder) of the new network have to be fine-tuned, requiring less labeled training data. ConvNets, transfer learning, and many other DL concepts have become building blocks of state-of-the-art animal tracking to address the specific needs in research settings.

State-of-the-Art Animal Motion Capture with Deep Learning

As for so many fields, DL revolutionized tracking of animals in various ways, from robustly tracking single (Mathis et al., 2018; Pereira et al., 2019) and multiple animals (Romero-Ferrero et al., 2019) and their poses (Pereira et al., 2020; Lauer et al., 2021) to 3D pose estimation in different species (Arac et al., 2019; Dickinson et al., 2020; Dunn et al., 2021). Generally, most current pose estimation frameworks, which are mainly in 2D, use ConvNets and an encoder-decoder architecture (see above), where the representation of input image in the encoder and key points (landmarks) of the pose in the decoder are jointly learned (Fig. 60–2D). Instead of landmarks, it is common practice to learn landmarks as as heat maps, so-called confidence maps (Carreira et al., 2016; Pereira et al., 2019), different density functions in the image where the brightest pixel of a particular confidence map represents the most likely location of a particular landmark.

Pose Estimation and the Dilemma with Animal Datasets

Directly learning to extract relevant features from images with DL usually requires large, labeled datasets for training, which is a problem in particular in laboratory settings with limited acquisition and labeling capacities. There are currently five ways how this labeling problem is addressed: transfer learning, efficient neural network design, active learning, marker-based ground truth, and synthetic datasets.

One of the most widely used animal pose estimation software package is DeepLabCut (Mathis et al., 2018), which takes advantage of transfer learning (see above) to lower requirements of labeled data and shorten training times. DeepLabCut pretrain their models on ImageNet, factually training an encoder for general-purpose visual feature detection (Mathis et al., 2018; Nath et al., 2019). However, to boost robustness on out-of-domain data (e.g., held out species during pretraining), DeepLabCut recently switched from this task-agnostic pretraining to a more domain-specific framework, pretraining their encoders directly for animal-pose-specific feature detection on a purposefully created SuperAnimal dataset, which provides uniform key points across species (Ye et al., 2021). Also capable of transfer learning is DeepBehavior (Arac et al., 2019), a tool box that offers different ConvNet architectures, which can be pretrained on datasets like CoCo (Lin et al., 2014) or ImageNet (Deng et al., 2009; Russakovsky et al., 2015) that offer bounding boxes annotation (i.e., imaginary rectangles outlining the object of interest in an image). One of the architectures in DeepBehavior is an improved version of the well-known model YOLO (“You Only Look Once”) (Redmon et al., 2016; Redmon and Farhadi, 2018), a fast ConvNet architecture that simultaneously predicts spatially separated bounding boxes and associated class probabilities to localize and categorize objects in images. Although YOLO models are elegantly designed for real-time object detection and are great in laboratory settings to quantify simple objects (e.g., in a three-chamber test) and social interactions (e.g., in the case of two mice with miniscopes) (Arac et al., 2019), their focus on predicting bounding boxes makes such models less applicable for accurately estimating more complex poses.

Another popular toolkit is LEAP (Pereira et al., 2019) and its successor DeepPoseKit (Graving et al., 2019), which designed efficient neural network with smaller architectures (e.g., less-deep—i.e., fewer layers—ConvNets) and thus need fewer parameters to be tuned when training on small labeled datasets. While neural networks like LEAP are primarily used on images of animals in constrained laboratory settings with more uniform imaging conditions, neural networks like DeepLabCut try to be more multipurpose, allowing pose estimation in both laboratory settings and “in the wild.”

Apart from such architectural choices, other approaches such as active learning and the related human-in-the-loop training seek to improve labels on smaller datasets by starting with a small number of images that capture the diversity of representative features in the dataset and then they iteratively let the network predict and the human observer correct labels on new data over several training loops (Collins et al., 2008; Branson et al., 2010). Others (Dunn et al., 2021) removed the burden of manual labeling entirely by combining different camera systems and synchronously acquiring training data (i.e., color images) with regular color cameras and corresponding ground truth labels with infrared cameras for marker-based motion capture (e.g., by using retroreflective piercings). While these datasets with marker-based motion capture itself can be used to study animal behavior (e.g., idiosyncratic perseverative grooming sequences in a rodent model of fragile X syndrome; Marshall et al., 2021), methods requiring markers to create ground truth datasets may not be a feasible option in a number of paradigms (e.g., for species like zebrafish—Danio rerio—or octopus). Hence, animating photorealistic 3D animal models in different poses with exact body part labels are a valuable alternative, with an opportunity to synthetically create a potentially infinite ground truth dataset for animal pose estimation (Bolaños et al., 2021). While all these different methods approach the general labeling problem in different ways, they show different strength and weaknesses, often depending on the recording conditions (e.g., pose estimation in 3D or for multiple animals).

3D Pose Estimation, Dense Tracking, and Beyond

Behavior, the temporal dynamics of poses (including postures and facial expressions), can be decomposed into movements of individual body parts and thus inherently takes place in three dimensions (3D). For 3D pose estimation in animals, most approaches combine two-dimensional (2D) ConvNets and traditional triangulation to retrieve the 3D positions of the detected 2D key points, including DeepLabCut (Nath et al., 2019), DeepBehavior (Arac et al., 2019), and Anipose (Dickinson et al., 2020) (Fig. 60–2E). Triangulation requires the cameras to be robustly calibrated beforehand, usually by deploying a calibration board (e.g., a checkerboard or AprilTags; Olson, 2011), and some use triangulation to reduce the labeling effort by projecting key points from one view to another (e.g., as used in the 62-camera setup of OpenMonkeyStudio; Bala et al., 2020). Calibration, the process of estimating camera internal (“intrinsics” like the focal length) and external (“extrinsics,” including rotation and translation) parameters, has a long history in computer vision with several highly accurate algorithms (e.g., bundle adjustment [Triggs et al., 2000] or direct linear transformation [Hartley and Zisserman, 2004]). Some 3D animal pose estimation frameworks (Zhang and Park, 2020) take advantage of the calibration and train ConvNets with supervision from the epipolar geometry, enforcing between camera views constraints on the 2D key point detection. In contrast, DeepFly3D (Günel et al., 2019) calibrates a multicamera setup without a calibration board and instead uses pictorial structures (Felzenszwalb and Huttenlocher, 2005) of detected key points in images, specifically line segments representing the legs of a tethered Drosophila, to improve through active learning iteratively the 3D pose accuracy of the fly’s legs. Leveraging the rigidness of rodent skeleton and the consistency of joint angles between similar postures, GIMBAL (Zhang et al., 2020) introduces spatiotemporal priors to their model to improve 3D key point detection. DAANCE (Dunn et al., 2021), on the other hand, uses a coarse estimate of each camera position and orientation to project features from 2D ConvNets into 3D space, where a 3D ConvNet fuses features from different views to predict accurate landmarks in 3D. While most approaches still require some form of precalculated triangulation for new data, LiftPose3D (Gosztolai et al., 2020) aimed to learn the nonlinear mapping between 2D and 3D poses directly in order to achieve high-quality 3D pose estimation in the absence of large camera arrays and calibration. As hopefully illustrated by this list of selected algorithms depicting the diversity of sophisticated architectural designs, the computer vision and animal tracking community has started to tackle several key limitations of DL-based 3D skeleton tracking in different laboratory settings, including the demand for specialized hardware (e.g., large arrays of calibrated cameras) and the robustness of DL-based methods compared to maker-based approaches.

While 3D pose estimation frameworks gain popularity in the neuroscience community, not the least due to their relevance in assessing motor impairments (Machado et al., 2015), there is a natural desire in neuroscience research community to describe the behavioral state of an animal beyond its pose, including somatosensory sensations like tactile information, and extend skeleton-based approaches toward a holistic 3D surface representation of the animal (Fig. 60–2F). Capturing the 3D surface of a body is a difficult task that receives growing research interest in computer vision. Although most research focuses on human bodies, there are a few attempts to apply the same techniques to animals; however, so far none of them has found its way into basic neuroscience research yet. Generally, there are two different approaches to capture the 3D surface of a body and enabling a dense pose tracking: model-based and reconstruction-based methods. Most model-based approaches leverage the success of key point and feature detection methods to establish correspondence between input data (e.g., an image depicting a person) and a 3D model (e.g., a generic human in a specific pose). Models are usually created using 3D scans, and there is a body of work in building large datasets of 3D models of humans in different poses and interactions with the environment, which includes the datasets and models of SCAPE (Anguelov et al., 2005), FAUST (Bogo et al., 2014), SMPL (Loper et al., 2015), Frank/Adam (Joo, Simon, and Sheikh, 2018), and GRAB (Taheri et al., 2020). Since such datasets build the foundation of almost all model-based approaches, similar datasets would need to be acquired for a variety of animals. Although these 3D setups are scaled and optimized to scan bodies of humans, there have been attempts to acquire animal datasets within similar setups, as shown for dogs (Kearney et al., 2020). To address the challenge of acquiring large datasets in a variety of animals, some of which being less cooperative (e.g., tigers and birds), 3D scans of animal figurines in different poses (e.g., SMAL [Zuffi et al., 2017]) or articulated 3D mesh models created by an artist (Badger et al., 2020) have shown to be great alternatives. Similar to pipelines of model-based approaches in humans, different DL approaches are deployed to then learn deforming these 3D animal-shaped models, so they fit the pose and shape of an animal in an image. Such model-based approach have already shown great success for dense pose estimation in a variety of animal image datasets and some smaller video datasets that include animals such as dogs (Biggs et al., 2019), birds (Badger et al., 2020; Y. Wang et al., 2021), tigers (Zuffi, Kanazawa, and Black, 2018), and zebras (Zuffi et al., 2019). Other methods that are in a broader sense model-based comprise approaches that try to learn the 3D shape of a particular class of animals directly from large image datasets of that class of animal (as shown in birds [Kanazawa et al., 2018] and dogs [Biggs et al., 2020]) or approaches based on DensePose (Güler, Neverova and Kokkinos, 2018) that directly learn the correspondence between key points on an image and a 3D surface model (as shown in chimpanzees [Sanakoyeu et al., 2020] and a variety of quadrupeds [Neverova et al., 2021]). However, mostly due to the lack of large 3D video datasets of real animals (as available for humans; see list of datasets above), model-based approaches for dense animal pose tracking to date are still struggling to create realistic movement patterns in 3D, with some research groups starting to integrate additional motion capture data to address this problem (e.g., horse-specific dataset called hSMAL [Li et al., 2021]). In contrast to model-based approaches, reconstruction-based methods are agnostic to the identity of objects and can reconstruct entire scenes. There are several approaches for 3D reconstruction with a long history in computer vision (Fitzgibbon and Zisserman, 1998; Seitz et al., 2006; Newcombe et al., 2011; Niessner et al., 2013; Dou et al., 2016; Schönberger et al., 2016; Innmann et al., 2020). While an individual description of these techniques is beyond the scope of this chapter, most of them rely on large sets of images that capture a scene from different viewpoints either using several regular RGB (red-green-blue) cameras or a few specialized RGB-D cameras that provide additional depth (“D”) information like stereo cameras (e.g., Intel® RealSense™) or time-of-flight (ToF) cameras (e.g., Microsoft® Kinect™). While some studies use multiple RGB-D cameras to facilitate the tracking of 3D poses (Matsumoto et al., 2013; Ebbesen and Froemke, 2020), to date there are no approaches for true 3D surface reconstruction that are used in neuroscience research settings. However, promising results come from studies using single RGB-D cameras and thus depth information (sometimes refer to as 2.5D) for behavioral tracking, which highlights the relevance of information that goes beyond simple poses and the necessity of incorporating details about an animal’s surface for behavior analysis. Such approaches were successfully used for unsupervised phenotyping of rodent behavior (Wiltschko et al., 2015, 2020), the study of striatal (Markowitz et al., 2018) and cerebellar (Rudolph et al., 2020) circuits, as well as studies focusing on complex social behaviors (Hong et al., 2015; de Chaumont et al., 2019).

Multianimal Pose Estimation

Multianimal tracking has attracted more attention recently, not the least because of initiatives in the neuroscience community to study animal behavior in more natural and complex environments (e.g., Dennis et al., 2021). However, due to close interactions between animals, pose estimation with multiple animals is challenging. It is a daunting task to detect key points (i.e., body parts) in video frames of interacting animals and assign them accurately to individuals across frames. As discussed and supported by the software package SLEAP (Pereira et al., 2020), there are two main strategies for pose estimation with multiple animals: top-down and bottom-up. In the top-down approach, every animal (“instance”) is first localized (“anchored”) in the image and then body parts are detected within cropped images of each instance. In the bottom-up approach, all body parts are first detected and then assigned to different instances. The top-down (“instances-then-parts”) approach uses two distinct neural networks, one for instance and one for body part detection, factually predicting separate confidence maps for each, and requiring running the second neural network multiple times (i.e., for each animal separately). In contrast, the bottom-up (“parts-then-instances”) approach uses only one neural network and predicts for all body parts confidence maps and so-called part affinity fields (Cao et al., 2017), 2D vector fields that encode the location and orientation of each body part, which can be used to connect body parts to directed graphs (“skeletons”) for each instance. DeepLabCut has also recently added multianimal pose estimation to their framework, testing their bottom-up approach in different species (mice, marmosets, and fish) and a dataset with parenting mice (Lauer et al., 2021). Together with recent successes in identifying and tracking multiple animals in large collectives of up to 100 animals (Romero-Ferrero et al., 2019), it seems just a question of time when pose dynamics of animals, their behavior, can be accurately measured in ethologically meaningful or stressful social settings (e.g., such as John Calhoun’s “behavioral sink”; Calhoun, 1962).

Quantifying Behavior

Reference Coordinates

Behavior is composed of coordinated movements of individual body parts and, as such, can be described by the animal’s pose at different levels of detail (see above and Fig. 60–2). Beside the granularity of the pose description, behavior can be analyzed agnostic to the surrounding environment (“egocentric representation”) or with respect to the surrounding environment (“allocentric representation”) (Fig. 60–1B), with some behavioral analysis software exploiting the first (like MotionMapper [Berman et al., 2014] and MoSeq [Wiltschko et al., 2015]) and others the second (JAABA [Kabra et al., 2013], B-SOiD [Hsu and Yttri, 2019], LiveMouseTracker [de Chaumont et al., 2019], SimBA [Nilsson et al., 2020], and MARS [Segalin et al., 2020]) for their analysis. Analyzing pose dynamics independently of position and orientation in the environment, poses in 2D or 3D can be centered at new egocentric coordinates and aligned by using two (e.g., head and pelvis) or three (e.g., plus sternum) anatomical landmarks, respectively. In this egocentric coordinate space, the movement of individual body parts or the relation between them can be quantified based on the speed and direction of the coordinate transformation of different anatomical landmarks (Fig. 60–1B). An “egocentric representation” of behavior can be used, for example, as kinematic description of both voluntary (i.e., physiological) behavior such as grooming and involuntary (i.e., pathological or optogenetically induced) behaviors such as “neomorphic” waddling or spinning (Wiltschko et al., 2015). For basic epilepsy research specifically, the egocentric representation of pose dynamics can replace labor-intensive, manual assessment of seizure behavior using the Racine scale (Racine, 1972) with an unbiased description of ictal behavior that can be acquired automatically within a variety of environmental contexts (e.g., home cage or different behavioral assays) (Fig. 60–1B). In contrast to representing poses in an egocentric coordinate system, velocity, distance, and angle between anatomical and environmental landmarks in the original allocentric coordinate system describes an animal’s interaction with its environment. Hence, an “allocentric representation” proves useful, for example, to assess higher order functions or dysfunctions in epileptic animals, such as spatial discrimination and memory functions (e.g., in the object location memory task [Bui et al., 2018]), or to infer an animal’s state of mind with regard to a particular environmental context (e.g., anxiety-related avoidance behavior in the open arms of an elevated plus maze). In essence, selectively switching between egocentric and allocentric representations of pose dynamics provides a way to translate the features tracked with DL approaches into classical descriptions of epileptic phenotypes and enables future AI-guided findings to be put in the context of the large body of work analyzing behavior in animal models of epilepsies with traditional approaches (Fig. 60–1).

Decomposing the Temporal Structure of Behavior

The temporal structure of behavior is often thought to be composed of a set of discrete, reoccurring stereotypic modules as shown in a variety of model organisms in neuroscience—Drosophila melanogaster (Berman et al., 2014; Calhoun, Pillow, and Murthy, 2019), Caenorhabditis elegans (Yemini et al., 2013; Linderman et al., 2019), Danio rerio (Marques et al., 2018; Johnson et al., 2020), Mus musculus (Wiltschko et al., 2015; Markowitz et al., 2018), and Rattus norvegicus (Marshall et al., 2021). Measures of behavior (e.g., video frames or tracked features) can be decomposed into (behavioral) modules by a variety of methods (see also reviews: Anderson and Perona, 2014; Brown and De Bivort, 2018; Datta et al., 2019; Pereira, Shaevitz and Murthy, 2020). While both raw data (e.g., video frames) or extracted behavioral features (e.g., animal’s pose) in either 2D or 3D can be used as input data (Fig. 60–4A), it is common to align them to egocentric coordinates first. Similar to the alignment of tracked poses (see above), raw video footage (or the 3D voxel-representation of a scene), which in addition captures nonrigid body movements (e.g., facial expressions [Stringer et al., 2019; Dolensek et al., 2020]), can be cropped to the size and aligned to the orientation of an animal (Fig. 60–4A), which than can be used as input data for behavioral segmentation (Fig. 60–4B).

Figure 60–4.. Decomposition and quantification of behavior.

Figure 60–4.

Decomposition and quantification of behavior. A. Input data to decompose behavior into discrete behavioral modules can be either the raw video image data or extracted behavioral features from this video data. In 2D, a sequence of either individual video (more...)

Generally, techniques that decompose the temporal structure of behavioral data into stereotypic, reoccurring segments, which are sometimes referred to as behavioral modules, can be roughly divided into two main categories: supervised and unsupervised methods. While there are a variety of different supervised ML approaches to segment behavior (e.g., decision trees or random forest [Kabra et al., 2013; Hong et al., 2015; Nilsson et al., 2020]), all supervised methods rely on a large set of labeled data, with the possible exception of self-supervised techniques (not discussed here). Besides pure manual labeling, one common practice to facilitate the creation of human-provided labels is to use the tracked features and a set of criteria. These predefined rules range from simply thresholding different tracked features of a single animal (e.g., speed threshold for the label “running”) to complex combinatorial rules for social descriptions (including dyadic dynamics and group-building behavior) using pose annotations from multiple animals (de Chaumont et al., 2012, 2019; Kabra et al., 2013; Segalin et al., 2020). While such algorithms can drastically reduce the annotation time, variation in human annotation (e.g., inter-annotator and inter-lab differences) remains a problem (Szigeti, Stone, and Webb, 2016; Leng et al., 2020; Segalin et al., 2020), which some researchers try to address directly by unveiling the annotation style itself (e.g., highlighting the relevance of behavioral features for an annotator’s choice of a label) to improve annotator consensus and thus reproducibility of behavioral studies (Tjandrasuwita et al., 2021).

Unsupervised methods, on the other hand, elude the requirement for human-provided labels entirely and provide a data-driven alternative without a priori assumptions about the feature composition of different behavioral modules. Most of these unsupervised pipelines first deploy linear (e.g., principal component analysis [PCA] [Jolliffe, 2002]) or nonlinear dimensionality reduction technique (e.g., manifold embedding like t-stochastic neighbor embedding, t-SNE [van der Maaten and Hinton, 2008]) to remove redundant information in the data that arises, for example, from correlative movements between body parts. For example, it was found that projecting high-dimensional posture descriptions onto a low-dimensional basis (i.e., by applying PCA), which is formed from eigenvectors that each represent a fundamental movement and/or pose, 95% of the shape variance in Caenorhabditis elegans can be described by just four dimensions (eigenvectors) (Stephens et al., 2008). These dimensions (or modes) were termed “eigenworms” (Stephens et al., 2008; Ahamed, Costa, and Stephens, 2021), and since then, similar “postural eigenmodes” were used in other animals such as fruit flies (Berman et al., 2014; Werkhoven et al., 2019), zebrafish (Girdhar, Gruebele, and Chemla, 2015), and rodents (Wiltschko et al., 2015; Marshall et al., 2021). After dimensionality reduction, there are two ways to analyze the preprocessed data: discretize low-dimensional data into (behavioral) modules or analyze its continuous representation. For the first one, there are two main strategies: clustering or modeling. Clustering algorithms such as k-means (Lloyd, 1982), Gaussian mixture models (McLachlan and Peel, 2000), or density-based clustering algorithms (e.g., watershed-based heuristic [Meyer, 1994; Berman et al., 2014] or DBSCAN [Ester et al., 1996]) will group similar points together with different strategies, using similarity metrics such as Euclidean distance (Fig. 60–4B). A popular behavioral analysis pipeline is MotionMapper (Berman et al., 2014), which deploys many of these techniques and is used in different animal models and experimental settings (Berman et al., 2014; Wang et al., 2016; Klibaite et al., 2017; Liu et al., 2018; Merel et al., 2019; Pereira et al., 2019; Zimmermann et al., 2020; Marshall et al., 2021). Specifically, MotionMapper uses PCA on aligned data, followed by converting low-dimensional time series into a frequency domain representation using a Morlet wavelet transform (Goupillaud, Grossmann, and Morlet, 1984), which is mapped with t-SNE to 2D and clustered with watershed transform into regions on the 2D embedding, where peaks represent more and valleys represent less stereotypic behaviors. Others like B-SOID (Hsu and Yttri, 2019) or OpenMonkeyStudio (Bala et al., 2020) are conceptually similar, but replace PCA with state-of-the-art dimensional reduction techniques like UMAP (i.e., uniform manifold proximation and projection) (McInnes, Healy, and Melville, 2018) and t-SNE with hierarchical DBSCAN (Campello, Moulavi, and Sander, 2013). In contrast to such clustering approaches, modeling the behavioral dynamics bears the potential to gain more insight into the structure of behavioral dynamics by reducing their complexity to a few model parameters. Motion sequencing (MoSeq), another popular behavioral analysis pipeline, uses a probabilistic graphical model, an autoregressive hidden Markov model (HMM), to identify hidden states and their transitions in low-dimensional (i.e. PCA preprocessed) RGB-D data (Wiltschko et al., 2015, 2020; Johnson et al., 2016; Pisanello et al., 2017; Markowitz et al., 2018; Rudolph et al., 2020). Noteworthy, MoSeq models each behavioral module as continuous auto-regressive process in behavioral-state space and thus builds a hybrid between analysis approaches with discretized and continuous representations of behavior. Generally, continuous representations take account of the continuous patterns of motion (e.g., oscillator structure during locomotion) and are used to analyze a variety of complex behaviors, including hunting behaviors of zebrafish and other stereotyped behaviors [Stephens et al., 2008, 2011; Bolton et al., 2019; DeAngelis, Zavatone-Veth, and Clark, 2019; Mearns et al., 2020; Ahamed, Costa and Stephens, 2021]). More recently, DL approaches such as structured VAE (Johnson et al., 2016) and VAE-SNE (Graving and Couzin, 2020) showed their great potential for extracting such continuous dynamic representations from behavioral time series in a flexible and reliable manner, while simultaneously being able to discretize them into stereotypic modules. Although there are still several open challenges about how to best segment continuous behavior into behavioral components and recover the hierarchical organization of behavior (Berman, Bialek, and Shaevitz, 2016; Datta et al., 2019; Tao et al., 2019), many of the unsupervised techniques discussed here (such as hierarchical DBSCAN or hierarchical HMM) are capable of unraveling the hierarchical structure of a behavioral dataset (Fig. 60–4B).

Conclusion and Future Directions for Basic Epilepsy Research

A key obstacle to the faster development of new therapies for the epilepsies is that rigorous preclinical epilepsy research typically requires labor-intensive and expensive 24/7 video-EEG monitoring followed by subjective scoring of behavioral seizures as exemplified by the Racine scale (Racine, 1972). While automated electrographic seizure detection algorithms are improving, the critically important behavioral manifestations of epilepsy both during ictal and interictal periods remain poorly quantified and are subject to observer bias. As discussed in this chapter, recent technological advances in AI-guided quantification of behavior, including animal motion tracking and description of behavioral dynamics, highlight a promising future for behavioral phenotyping in basic epilepsy research. While the techniques discussed in this chapter are not yet adopted broadly by the epilepsy community, we would like to highlight two main areas where AI-guided phenotyping in epilepsy would be particularly impactful: for screening and mechanistic insights.

AI-Guided Phenotyping in Epilepsy for Screening at Scale

Behavioral phenotyping has a long history in epilepsy research and is well described in a variety of animal models, including fruit flies (Parker et al., 2011) and zebrafish (Baraban et al., 2005; Griffin et al., 2017, 2021). In rodent models of epilepsy, a large body of work showed that a spectrum of behavioral symptoms can be correlated with electrographic discharges, which was used to create a variety of different behavioral seizure scales (Racine, 1972; Jobe, Picchioni, and Chin, 1973; Pinel and Rovner, 1978a, 1978b; Pohl and Mares, 1987; Haas, Sperber, and Moshe, 1990; Veliskova et al., 1990). The symptoms observed range from behavioral arrest and automatisms to wild running and tonic-clonic movements of the limbs. While the changes with strong behavioral components are easier to identify, other, more subtle changes (e.g., staring, trembling of whiskers) require a trained eye and therefore are highly dependent on the human observer. As an objective alternative to the Racine and other scales, AI-guided phenotyping is an unbiased approach to partitioning and labeling complex behaviors, providing an opportunity to systematically characterize epileptic phenotypes beyond traditional behavioral classification terms (i.e., hand-labeling coarse behavioral classes, as in the Racine scale [Racine, 1972]), automatically quantify epileptic seizures, and search for stereotypic behavioral modules and transitions that are not a priori defined to capture potentially unrecognized epileptic phenotypes. For example, it was shown that AI-guided behavioral analysis can identify subtle behaviors that are normally not distinguished by a human observer as, for example, for the waddling gait in Ror1b mutant mice (Wiltschko et al., 2015). Moreover, AI-guided behavioral phenotyping has a great potential for anti-seizure drug (ASD) screening at scale. For example, Wiltschko et al. (2020) showed that AI-guided behavioral phenotyping with the analysis pipeline MoSeq is capable of automatically discriminate between mice injected with different neuroactive and psychoactive drugs and can even be used to identify specific on- and off-target effects of drugs in a mouse model of autism spectrum disorder. One could imagine that a similar approach—an out-of-the-box methodology for AI-guided behavioral phenotyping in epilepsy, offering a powerful, freely shared experimental and analytical tool for quantifying complex behavior with subsecond precision—would have great impact on basic epilepsy research. Such an approach would increase reliability of behavioral phenotyping and accelerate the assessment of both established and candidate therapeutics in a variety of acquired and genetic epilepsy models. With growing efforts to introduce data standards such as Neurodata Without Borders (NWB) (Teeters et al., 2015; Rübel et al., 2019), one could envision that behavioral datasets similar to those of the International Brain Laboratory (Abbott et al., 2017; Bonacchi et al., 2019) with standardized descriptions for behavioral phenotypes could be shared across labs (e.g., in a central database or through individually managed datasets [Sun et al., 2021]), cross-examined, and used for generating new hypotheses about the mechanistic basis and potential treatment options for different phenotypes. For example, a large, publicly available dataset would open opportunities to exploit techniques such as transfer learning, if combined with behavioral analysis, which would be especially useful for rarer types of epilepsy syndromes or generally any phenotype with epilepsy-like symptoms where there is less data available such as organophosphate-induced epileptic phenotypes (Enderlin et al., 2020; Guignet et al., 2020) or radiation-induced memory deficits that are in many ways similar to the cognitive comorbidities observed in experimental TLE (Klein et al., 2021).

Linking Brain to Behavior in Epilepsy with AI-Guided Phenotyping

Leveraging the recent advance made in both AI-guided behavioral phenotyping (discussed here) and neurotechnologies (Vázquez-Guardado et al., 2020) will likely enable studies aiming to provide better mechanistic insights into alterations in neuronal activity underlying different motor and cognitive symptoms in epilepsy. The necessity to study the “Brain In Action” (Mott, Gordon and Koroshetz, 2018), a research priority of the NIH BRAIN Initiative (Bargmann and Newsome, 2014), applies to both physiological and pathological function. As pointed out by Datta et al. (2019), some of the most exciting discoveries in neuroscience in the past 50 years, which includes place cells (O’Keefe and Dostrovsky, 1971; O’Keefe, 1976), grid cells (Hafting et al., 2005) and replay (Skaggs and McNaughton, 1996; Nádasdy et al., 1999), were made in studies using neural recordings in freely behaving animals. Hence, studying their dysfunction in the epilepsies will also require both neural recordings and behavioral measures in more naturalistic and complex environments. For example, recent findings in animal models of TLE that show altered CA1 place cell (Liu et al., 2003; Shuman et al., 2020) and mossy cell function (Bui et al., 2018) indicate that more sophisticated exploration and learning paradigms (e.g., a more naturalistic and complex maze environment [Rosenberg et al., 2021]) will be helpful for gaining a better mechanistic understanding of spatial memory impairments in TLE. Navigation, as most behaviors, is complex and requires the integration of multiple cues through multimodal (e.g., visual or auditory) perception. Analyzing and modeling these modalities (Fig. 60–5A,B) can be done with similar approaches as discussed above for behavior in general. For example, there is a rich literature of different unsupervised clustering and modeling techniques that can be used to analyze acoustic events in animals (Pearre et al., 2017; Coffey, Marx and Neumaier, 2019; Sainburg, Thielk and Gentner, 2020; Fonseca et al., 2021), and these insights have shown to be useful in particular for social assays, where ultrasonic vocalizations (USVs) seem to be associated with distinct behaviors (Sangiamo, Warren and Neunuebel, 2020). Interestingly, ancillary information such as acoustic data is sometimes necessary to assist behavioral analysis pipelines in distinguishing similar behaviors and inferring motivational states, as shown in mice for same- and opposite-sex mounting behavior that could only distinguish between sexual and aggressive motivation by the presence and absence of USVs, respectively (Karigo et al., 2021). Altogether, such multimodal measurements can be combined in so-called agent-based models, where an artificial neural network (“agent”) can be trained to react based on perceived stimuli in a virtual environment. These agent-based models have recently gained more attention in neuroscience, including building “playgrounds” for cognitive AI in form of classical animal cognition tasks (Beyret et al., 2019; Crosby, Beyret and Halina, 2019) or modeling behavioral strategies and kinematics in a virtual rodent in an neuroethologically meaningful manner (Merel et al., 2019). Interestingly, training such agent-based, abstract models like a recurrent neural network for path integration lead to the emergence of grid-like representations similar to those seen in the entorhinal cortex (Banino et al., 2018). Such computational approaches may prove to be particularly useful for modeling alterations in cognitive functions in basic epilepsy research (e.g., spatial memory deficits), as they can use in silico perturbation strategies similar to those performed in biophysical models for studying the neuropathophysiology of the epilepsies (Case et al., 2012). Beyond the field of navigation research, agent-based models may also become increasingly attractive for modeling sensorimotor systems in combination with marker-less motion capture and biomechanical modeling (Merel et al., 2019; Sandbrink et al., 2020; Hausmann et al., 2021). In animal models of epilepsy, such combined approaches of acquiring and modeling data of sensorymotor systems would be highly valuable for developing a better understanding of the semiologies that are associated with distinct seizure types.

Figure 60–5.. A possible future of hypothesis-driven research in epilepsy.

Figure 60–5.

A possible future of hypothesis-driven research in epilepsy. A. Behavioral experiments such as object location memory tasks usually consist of task-related environmental stimuli and include multimodal measurements (e.g., video and audio recordings) that (more...)

In summary, AI-guided phenotyping provides an excellent interface between experimental and computational epilepsy research. On the one hand, AI-guided tracking algorithms provide a more functionally relevant (i.e., behavioral) readout that can be correlated to or even jointly modeled (Batty et al., 2019) with neural activity (Fig. 60–5A,C). On the other hand, AI-guided behavioral phenotyping can provide training data (e.g., field of view, pose, and movement trajectory) for computational models such as agent-based models to run virtual experiments on the basis of behavioral data from epileptic animals (Fig. 60–5B). The latter has potential to inspire new hypotheses that can be tested experimentally (e.g., with real-time closed-loop optogenetics [Armstrong et al., 2013; Krook-Magnuson et al., 2013]) (Fig. 60–5C), and thus will likely be of increasingly greater benefit for quantitative, rigorous hypothesis-driven research in the epilepsies.

Acknowledgments

Grant support to IS from the National Institutes of Health (5R01NS114020-02) and to TG from the Swiss National Science Foundation (174811 and 186757) is gratefully acknowledged. TG would also like to thank Frances Cho for useful discussions and constructive criticism.

Disclosure Statement

The authors declare no relevant conflicts.

References

  1. Abbott, L. F. et al. (2017) ‘An International Laboratory for Systems and Computational Neuroscience’, Neuron, 96(6), pp. 1213–1218. doi: 10.1016/j.neuron.2017.12.013. [PMC free article: PMC5752703] [PubMed: 29268092]
  2. Abrahams, B. S. et al. (2011) ‘Absence of CNTNAP2 Leads to Epilepsy, Neuronal Migration Abnormalities, and Core Autism-Related Deficits’, Cell, 147(1), pp. 235–246. doi: 10.1016/j.cell.2011.08.040. [PMC free article: PMC3390029] [PubMed: 21962519]
  3. Achanta, R. et al. (2010) ‘SLIC Superpixels’, EPFL Technical Report 149300, (June).
  4. Ahamed, T., Costa, A. C. and Stephens, G. J. (2021) ‘Capturing the continuous complexity of behaviour in Caenorhabditis elegans’, Nature Physics, 17(2), pp. 275–283. doi: 10.1038/s41567-020-01036-8.
  5. Anderson, D. J. and Perona, P. (2014) ‘Toward a science of computational ethology’, Neuron, 84(1), pp. 18–31. doi: 10.1016/j.neuron.2014.09.005. [PubMed: 25277452]
  6. Anguelov, D. et al. (2005) ‘SCAPE: Shape Completion and Animation of People’, ACM Trans. Graph., 24(3), pp. 408–416. doi: 10.1145/1073204.1073207.
  7. Arac, A. et al. (2019) ‘Deepbehavior: A deep learning toolbox for automated analysis of animal and human behavior imaging data’, Frontiers in Systems Neuroscience, 13(May), pp. 1–12. doi: 10.3389/fnsys.2019.00020. [PMC free article: PMC6513883] [PubMed: 31133826]
  8. Armstrong, C. et al. (2013) ‘Closed-loop optogenetic intervention in mice’, Nat Protoc. 2013/07/11, 8(8), pp. 1475–1493. doi: 10.1038/nprot.2013.080. [PMC free article: PMC3988315] [PubMed: 23845961]
  9. Badger, M. et al. (2020) ‘3D Bird Reconstruction: A Dataset, Model, and Shape Recovery from a Single View’, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 12363 LNCS, pp. 1–17. doi: 10.1007/978-3-030-58523-5_1. [PMC free article: PMC9273110] [PubMed: 35822859]
  10. Bala, P. C. et al. (2020) ‘Automated markerless pose estimation in freely moving macaques with OpenMonkeyStudio’, Nature Communications, 11(1), pp. 1–12. doi: 10.1038/s41467-020-18441-5. [PMC free article: PMC7486906] [PubMed: 32917899]
  11. Banino, A. et al. (2018) ‘Vector-based navigation using grid-like representations in artificial agents’, Nature, 26, p. 1. Available at: http://www​.nature.com​/articles/s41586-018-0102-6%0Ahttp:​//dx​.doi.org/10.1038/s41586-018-0102-6. [PubMed: 29743670]
  12. Baraban, S. C. et al. (2005) ‘Pentylenetetrazole induced changes in zebrafish behavior, neural activity and c-fos expression’, Neuroscience, 131(3), pp. 759–768. doi: 10.1016/j.neuroscience.2004.11.031. [PubMed: 15730879]
  13. Bargmann, C. I. and Newsome, W. T. (2014) ‘The brain research through advancing innovative neurotechnologies (brain) initiative and neurology’, JAMA Neurology, 71(6), pp. 675–676. doi: 10.1001/jamaneurol.2014.411. [PubMed: 24711071]
  14. Batty, E. et al. (2019) ‘BehaveNet: Nonlinear embedding and Bayesian neural decoding of behavioral videos’, Advances in Neural Information Processing Systems, 32(NeurIPS 2019), pp. 15706-15717.
  15. Bengio, Y. et al. (2011) ‘Deep learners benefit more from out-of-distribution examples’, Journal of Machine Learning Research, 15, pp. 164–172.
  16. Bengio, Y. (2011) ‘Deep Learning of Representations for Unsupervised and Transfer Learning’, JMLR: Workshop and Conference Proceedings, 7, pp. 1–20.
  17. Berman, G. J. et al. (2014) ‘Mapping the stereotyped behaviour of freely moving fruit flies’, Journal of the Royal Society Interface, 11(99). doi: 10.1098/rsif.2014.0672. [PMC free article: PMC4233753] [PubMed: 25142523]
  18. Berman, G. J., Bialek, W. and Shaevitz, J. W. (2016) ‘Predictability and hierarchy in Drosophila behavior’, Proceedings of the National Academy of Sciences of the United States of America, 113(42), pp. 11943–11948. doi: 10.1073/pnas.1607601113. [PMC free article: PMC5081631] [PubMed: 27702892]
  19. Beyret, B. et al. (2019) ‘The Animal-AI Environment: Training and Testing Animal-Like Artificial Cognition’. Available at: http://arxiv​.org/abs/1909.07483.
  20. Biggs, B. et al. (2019) ‘Creatures Great and SMAL: Recovering the Shape and Motion of Animals from Video’, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 11365 LNCS, pp. 3–19. doi: 10.1007/978-3-030-20873-8_1.
  21. Biggs, B. et al. (2020) ‘Who Left the Dogs Out? 3D Animal Reconstruction with Expectation Maximization in the Loop’, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 12356 LNCS, pp. 195–211. doi: 10.1007/978-3-030-58621-8_12.
  22. Binder, D. K. et al. (2020) ‘Epilepsy Benchmarks Area II: Prevent Epilepsy and Its Progression’, Epilepsy Currents, 20(1_suppl), pp. 14S–22S. doi: 10.1177/1535759719895274. [PMC free article: PMC7031802] [PubMed: 31937124]
  23. Bogo, F. et al. (2014) ‘FAUST: Dataset and evaluation for 3D mesh registration’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3794–3801. doi: 10.1109/CVPR.2014.491.
  24. Bolaños, L. A. et al. (2021) ‘A three-dimensional virtual mouse generates synthetic training data for behavioral analysis’, Nature Methods, 18(4), pp. 378–381. doi: 10.1038/s41592-021-01103-9. [PMC free article: PMC8034498] [PubMed: 33820989]
  25. Bolton, A. D. et al. (2019) ‘Elements of a stochastic 1 3D prediction engine in larval zebrafish prey capture’, eLife, 8, pp. 1–24. doi: 10.7554/eLife.51975. [PMC free article: PMC6930116] [PubMed: 31769753]
  26. Bonacchi, N. et al. (2019) ‘Data architecture for a large-scale neuroscience collaboration’, bioRxiv, pp. 1–21. doi: 10.1101/827873.
  27. Bottou, L., Curtis, F. E. and Nocedal, J. (2018) ‘Optimization methods for large-scale machine learning’, SIAM Review, 60(2), pp. 223–311. doi: 10.1137/16M1080173.
  28. Branson, K. et al. (2009) ‘High-throughput ethomics in large groups of Drosophila’, Nature Methods, 6(6), pp. 451–457. doi: 10.1038/nmeth.1328. [PMC free article: PMC2734963] [PubMed: 19412169]
  29. Branson, S. et al. (2010) ‘Visual Recognition with Humans in the Loop’, in Daniilidis, K., Maragos, P., and Paragios, N. (eds) Computer Vision—ECCV 2010. Berlin, Heidelberg: Springer Berlin Heidelberg, pp. 438–451.
  30. Brown, A. E. X. and De Bivort, B. (2018) ‘Ethology as a physical science’, Nature Physics, 14(7), pp. 653–657. doi: 10.1038/s41567-018-0093-0.
  31. Bui, A. D. et al. (2018) ‘Dentate gyrus mossy cells control spontaneous convulsive seizures and spatial memory’, Science. 2018/02/17, 359(6377), pp. 787–790. doi: 10.1126/science.aan4074. [PMC free article: PMC6040648] [PubMed: 29449490]
  32. Calhoun, A. J., Pillow, J. W. and Murthy, M. (2019) ‘Unsupervised identification of the internal states that shape natural behavior’, Nature Neuroscience, 22(12), pp. 2040–2049. doi: 10.1038/s41593-019-0533-x. [PMC free article: PMC7819718] [PubMed: 31768056]
  33. Calhoun, J. B. (1962) ‘Population density and social pathology.’, Scientific American, 206, pp. 139–148. doi: 10.1038/scientificamerican0262-139. [PubMed: 13875732]
  34. Campello, R. J. G. B., Moulavi, D. and Sander, J. (2013) ‘Density-based clustering based on hierarchical density estimates’, in  Pei, J. et al. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2013. Lecture Notes in Computer Science, vol 7819. Springer, Berlin, Heidelberg. https://doi​.org/10.1007​/978-3-642-37456-2_14.
  35. Cao, Z. et al. (2017) ‘Realtime multi-person 2D pose estimation using part affinity fields’, Proceedings—30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, 2017-Janua, pp. 1302–1310. doi: 10.1109/CVPR.2017.143.
  36. Carreira, J. et al. (2016) ‘Human pose estimation with iterative error feedback’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2016-Decem, pp. 4733–4742. doi: 10.1109/CVPR.2016.512.
  37. Caruana, R. (1994) ‘Learning Many Related Tasks at the Same Time With Backpropagation’, NIPS’94: Proceedings of the 7th International Conference on Neural Information Processing Systems, 7, pp. 657–664.
  38. Case, M. J. et al. (2012) ‘Computer Modeling of Epilepsy.’, in Noebels, J. L. et al. (eds). Jasper’s basic mechanisms of the epilepsies, 4th edn. Oxford, New York, pp.298–311.
  39. Chang, B. S. et al. (2020) ‘Epilepsy Benchmarks Area I: Understanding the Causes of the Epilepsies and Epilepsy-Related Neurologic, Psychiatric, and Somatic Conditions’, Epilepsy Currents, 20(1_suppl), pp. 5S–13S. doi: 10.1177/1535759719895280. [PMC free article: PMC7031801] [PubMed: 31965828]
  40. de Chaumont, F. et al. (2012) ‘Computerized video analysis of social interactions in mice’, Nature Methods, 9(4), pp. 410–417. doi: 10.1038/nmeth.1924. [PubMed: 22388289]
  41. de Chaumont, F. et al. (2019) ‘Real-time analysis of the behaviour of groups of mice via a depth-sensing camera and machine learning’, Nature Biomedical Engineering, 3(11), pp. 930–942. doi: 10.1038/s41551-019-0396-1. [PubMed: 31110290]
  42. Chen, L. C. et al. (2018) ‘DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs’, IEEE Transactions on Pattern Analysis and Machine Intelligence, 40(4), pp. 834–848. doi: 10.1109/TPAMI.2017.2699184. [PubMed: 28463186]
  43. Coffey, K. R., Marx, R. G. and Neumaier, J. F. (2019) ‘DeepSqueak: a deep learning-based system for detection and analysis of ultrasonic vocalizations’, Neuropsychopharmacology, 44(5), pp. 859–868. doi: 10.1038/s41386-018-0303-6. [PMC free article: PMC6461910] [PubMed: 30610191]
  44. Collins, B. et al. (2008) ‘Towards Scalable Dataset Construction: An Active Learning Approach’, in  Forsyth, D., Torr, P., and Zisserman, A. (eds) Computer Vision—ECCV 2008. Berlin, Heidelberg: Springer Berlin Heidelberg, pp. 86–98.
  45. Crosby, M., Beyret, B. and Halina, M. (2019) ‘The Animal-AI Olympics’, Nature Machine Intelligence, 1(5), p. 257. doi: 10.1038/s42256-019-0050-3.
  46. Dalal, N. and Triggs, B. (2005) ‘Histograms of oriented gradients for human detection’, Proceedings—2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2005, Vol. 1, pp. 886–893. doi: 10.1109/CVPR.2005.177.
  47. Dankert, H. et al. (2009) ‘Automated monitoring and analysis of social behavior in Drosophila’, Nature Methods, 6(4), pp. 297–303. doi: 10.1038/nmeth.1310. [PMC free article: PMC2679418] [PubMed: 19270697]
  48. Datta, S. R. et al. (2019) ‘Computational Neuroethology: A Call to Action’, Neuron. 104(1), pp. 11–24. doi: 10.1016/j.neuron.2019.09.038. [PMC free article: PMC6981239] [PubMed: 31600508]
  49. DeAngelis, B. D., Zavatone-Veth, J. A. and Clark, D. A. (2019) ‘The manifold structure of limb coordination in walking Drosophila’, eLife, 8, pp. 1–33. doi: 10.7554/eLife.46409. [PMC free article: PMC6598772] [PubMed: 31250807]
  50. Dennis, E. J. et al. (2021) ‘Systems Neuroscience of Natural Behaviors in Rodents’, Journal of Neuroscience, 41(5), pp. 911–919. [PMC free article: PMC7880287] [PubMed: 33443081]
  51. Dickinson, E. S. et al. (2020) ‘Anipose : a Toolkit for Robust Markerless’. [PMC free article: PMC8498918] [PubMed: 34592148]
  52. Dolensek, N. et al. (2020) ‘Facial expressions of emotion states and their neuronal correlates in mice’, Science, 368(6486), pp. 89–94. doi: 10.1126/science.aaz9468. [PubMed: 32241948]
  53. Dou, M. et al. (2016) ‘Fusion4D: Real-time performance capture of challenging scenes’, ACM Transactions on Graphics, 35(4), pp. 1–13. doi: 10.1145/2897824.2925969.
  54. Dunn, T. W. et al. (2021) ‘Geometric deep learning enables 3D kinematic profiling across species and environments’, Nature Methods, 18(5), pp. 564–573. doi: 10.1038/s41592-021-01106-6. [PMC free article: PMC8530226] [PubMed: 33875887]
  55. Ebbesen, C. and Froemke, R. (2020) ‘Automatic tracking of mouse social posture dynamics by 3D videography, deep learning and GPU-accelerated robust optimization’. doi: 10.1101/2020.05.21.109629.
  56. Enderlin, J. et al. (2020) ‘Characterization of organophosphate-induced brain injuries in a convulsive mouse model of diisopropylfluorophosphate exposure’, Epilepsia, 61(6), pp. e54–e59. doi: 10.1111/epi.16516. [PubMed: 32359085]
  57. Ester, M. et al. (1996) ‘A density-based algorithm for discovering clusters in large spatial databases with noise’, in. AAAI Press, pp. 226–231.
  58. Farabet, C. et al. (2013) ‘Learning Hierarchical Features for Scene Labeling’, Pattern Analysis and Machine Intelligence, IEEE Transactions on, 35(8), pp. 1915–1929. doi: 10.1109/TPAMI.2012.231. [PubMed: 23787344]
  59. Felleman, D. J. and Van Essen, D. C. (1991) ‘Distributed hierarchical processing in the primate cerebral cortex’, Cerebral Cortex, 1(1), pp. 1–47. doi: 10.1093/cercor/1.1.1-a. [PubMed: 1822724]
  60. Felzenszwalb, P. F. and Huttenlocher, D. P. (2005) ‘Pictorial Structures for Object Recognition’, 61(1), pp. 55–79.
  61. Fisher, R. S. et al. (2017) ‘Operational classification of seizure types by the International League Against Epilepsy: Position Paper of the ILAE Commission for Classification and Terminology’, Epilepsia. 2017/03/10, 58(4), pp. 522–530. doi: 10.1111/epi.13670. [PubMed: 28276060]
  62. Fitzgibbon, A. and Zisserman, A. (1998) ‘Automatic Camera Recovery for Closed or Open Image Sequences’, in  ECCV.
  63. Fonseca, A. H. O. et al. (2021) ‘Analysis of ultrasonic vocalizations from mice using computer vision and machine learning’, eLife, 10, pp. 1–22. doi: 10.7554/eLife.59161. [PMC free article: PMC8057810] [PubMed: 33787490]
  64. Girdhar, K., Gruebele, M. and Chemla, Y. R. (2015) ‘The behavioral space of zebrafish locomotion and its neural network analog’, PLoS ONE, 10(7), pp. 1–18. doi: 10.1371/journal.pone.0128668. [PMC free article: PMC4489106] [PubMed: 26132396]
  65. Girshick, R. et al. (2014) ‘Rich feature hierarchies for accurate object detection and semantic segmentation’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 580–587. doi: 10.1109/CVPR.2014.81.
  66. Gomez-Marin, A. and Ghazanfar, A. A. (2019) ‘The Life of Behavior’, Neuron. 104(1), pp. 25–36. doi: 10.1016/j.neuron.2019.09.017. [PMC free article: PMC6873815] [PubMed: 31600513]
  67. Goodfellow, I., Bengio, Y. and Courville, A. (2016) Deep Learning. MIT Press.
  68. Gosztolai, A. et al. (2020) ‘LiftPose3D, a deep learning-based approach for transforming 2D to 3D pose in laboratory animals’, bioRxiv. doi: 10.1101/2020.09.18.292680. [PMC free article: PMC7611544] [PubMed: 34354294]
  69. Goupillaud, P., Grossmann, A. and Morlet, J. (1984) ‘Cycle-octave and related transforms in seismic signal analysis’, Geoexploration, 23(1), pp. 85–102. doi: 10.1016/0016-7142(84)90025-5.
  70. Graving, J. and Couzin, I. (2020) ‘VAE-SNE: a deep generative model for simultaneous dimensionality reduction and clustering’. bioRxiv, doi: 10.1101/2020.07.17.207993.
  71. Graving, J. M. et al. (2019) ‘Deepposekit, a software toolkit for fast and robust animal pose estimation using deep learning’, eLife, 8, pp. 1–42. doi: 10.7554/eLife.47994. [PMC free article: PMC6897514] [PubMed: 31570119]
  72. Griffin, A. et al. (2017) ‘Clemizole and modulators of serotonin signalling suppress seizures in Dravet syndrome’, Brain, 140(3), pp. 669–683. doi: 10.1093/brain/aww342. [PMC free article: PMC6075536] [PubMed: 28073790]
  73. Griffin, A. et al. (2021) ‘Phenotypic analysis of catastrophic childhood epilepsy genes’, Communications Biology, 4(1). doi: 10.1038/s42003-021-02221-y. [PMC free article: PMC8175701] [PubMed: 34083748]
  74. Guignet, M. et al. (2020) ‘Persistent behavior deficits, neuroinflammation, and oxidative stress in a rat model of acute organophosphate intoxication’, Neurobiology of Disease, 133(December 2018), p. 104431. doi: 10.1016/j.nbd.2019.03.019. [PMC free article: PMC6754818] [PubMed: 30905768]
  75. Güler, R. A., Neverova, N. and Kokkinos, I. (2018) ‘DensePose: Dense Human Pose Estimation In The Wild’, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7297–7306. Available at: http://arxiv​.org/abs/1612.01202.
  76. Günel, S. et al. (2019) ‘Deepfly3D, a deep learning-based approach for 3D limb and appendage tracking in tethered, adult Drosophila’, eLife, 8, pp. 1–23. doi: 10.7554/eLife.48571. [PMC free article: PMC6828327] [PubMed: 31584428]
  77. Haas, K. Z., Sperber, E. F. and Moshe, S. L. (1990) ‘Kindling in developing animals: expression of severe seizures and enhanced development of bilateral foci’, Brain Res Dev Brain Res. 1990/11/01, 56(2), pp. 275–280. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/2261687. [PubMed: 2261687]
  78. Hafting, T. et al. (2005) ‘Microstructure of a spatial map in the entorhinal cortex’, Nature, 436(7052), pp. 801–806. doi: 10.1038/nature03721. [PubMed: 15965463]
  79. Haji Maghsoudi, O., Vahedipour, A. and Spence, A. (2019) ‘A novel method for robust markerless tracking of rodent paws in 3D’, Eurasip Journal on Image and Video Processing, 2019(1). pp.1–19, doi: 10.1186/s13640-019-0477-9.
  80. Hartley, R. and Zisserman, A. (2004) Multiple View Geometry in Computer Vision. 2nd edn. Cambridge: Cambridge University Press. doi: DOI: 10.1017/CBO9780511811685.
  81. Hausmann, S. B. et al. (2021) ‘Measuring and modeling the motor system with machine learning’, Current Opinion in Neurobiology. 70, pp. 11–23. doi: 10.1016/j.conb.2021.04.004. [PubMed: 34116423]
  82. Hong, W. et al. (2015) ‘Automated measurement of mouse social behaviors using depth sensing, video tracking, and machine learning’, Proceedings of the National Academy of Sciences of the United States of America, 112(38), pp. E5351–E5360. doi: 10.1073/pnas.1515982112. [PMC free article: PMC4586844] [PubMed: 26354123]
  83. Horn, B. K. and Schunck, B. G. (1981) ‘Determining Optical Flow’, Artificial intelligence, 17(1–3), pp.185–203.
  84. Hsu, A. I. and Yttri, E. A. (2019) ‘B-SOiD: An open source unsupervised algorithm for discovery of spontaneous behaviors’, bioRxiv, pp. 1–40. doi: 10.1101/ 770271.
  85. Hubel, D. H. and Wiesel, T. N. (1962) ‘Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex’, The Journal of Physiology, 160(1), pp. 106–154. doi: 10.1113/jphysiol.1962.sp006837. [PMC free article: PMC1359523] [PubMed: 14449617]
  86. Hur, J. and Roth, S. (2020) ‘Optical Flow Estimation in the Deep Learning Age’, In: Noceti, N., Sciutti, A., Rea, F. (eds) Modelling Human Motion. Springer, Cham. pp. 119–140. doi: 10.1007/978-3-030-46732-6_7.
  87. Inayat, S. et al. (2020) ‘A matlab-based toolbox for characterizing behavior of rodents engaged in string-pulling’, eLife, 9, pp. 1–31. doi: 10.7554/eLife.54540. [PMC free article: PMC7347385] [PubMed: 32589141]
  88. Innmann, M. et al. (2020) ‘NRMVS: Non-Rigid Multi-view Stereo’, in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV).
  89. Deng et al. (2009) ‘ImageNet: A large-scale hierarchical image database’, In 2009 IEEE conference on computer vision and pattern recognition, pp. 248–255. Ieee. doi: 10.1109/cvprw.2009.5206848.
  90. Jobe, P. C., Picchioni, A. L. and Chin, L. (1973) ‘Role of brain norepinephrine in audiogenic seizure in the rat’, J Pharmacol Exp Ther. 1973/01/01, 184(1), pp. 1–10. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/4686009. [PubMed: 4686009]
  91. Johnson, M. J. et al. (2016) ‘Composing graphical models with neural networks for structured representations and fast inference’, Advances in Neural Information Processing Systems, 29, pp. 2954–2962.
  92. Johnson, R. E. et al. (2020) ‘Probabilistic Models of Larval Zebrafish Behavior Reveal Structure on Many Scales’, Current Biology, 30(1), pp. 70–82.e4. doi: 10.1016/j.cub.2019.11.026. [PMC free article: PMC6958995] [PubMed: 31866367]
  93. Jolliffe, I. T. (2002) Principal Component Analysis, Principal Component Analysis. New York: Springer-Verlag (Springer Series in Statistics). doi: 10.1007/b98835.
  94. Jones, J. E. et al. (2020) ‘Epilepsy Benchmarks Area IV: Limit or Prevent Adverse Consequence of Seizures and Their Treatment Across the Life Span’, Epilepsy Currents, 20(1_suppl), pp. 31S–39S. doi: 10.1177/1535759719895277. [PMC free article: PMC7031803] [PubMed: 31973592]
  95. Joo, H., Simon, T. and Sheikh, Y. (2018) ‘Total Capture: A 3D Deformation Model for Tracking Faces, Hands, and Bodies’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 8320–8329. doi: 10.1109/CVPR.2018.00868.
  96. Kabra, M. et al. (2013) ‘JAABA: Interactive machine learning for automatic annotation of animal behavior’, Nature Methods, 10(1), pp. 64–67. doi: 10.1038/nmeth.2281. [PubMed: 23202433]
  97. Kanazawa, A. et al. (2018) ‘Learning Category-Specific Mesh Reconstruction’, ArXiv, pp. 1–20.
  98. Karigo, T. et al. (2021) ‘Distinct hypothalamic control of same- and opposite-sex mounting behaviour in mice’, Nature, 589(7841), pp.258–263. doi: 10.1038/s41586-020-2995-0. [PMC free article: PMC7899581] [PubMed: 33268894]
  99. Kearney, S. et al. (2020) ‘RGBD-Dog: Predicting Canine Pose from RGBD Sensors’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 8333–8342. doi: 10.1109/CVPR42600.2020.00836.
  100. Khan, Z. et al. (2004) ‘Abnormal motor behavior and vestibular dysfunction in the stargazer mouse mutant’, Neuroscience, 127(3), pp. 785–796. doi: 10.1016/j.neuroscience.2004.05.052. [PubMed: 15283975]
  101. Kim, H. K. et al. (2020) ‘Optogenetic intervention of seizures improves spatial memory in a mouse model of chronic temporal lobe epilepsy’, Epilepsia, 61(3), pp. 561–571. doi: 10.1111/epi.16445. [PMC free article: PMC7708390] [PubMed: 32072628]
  102. Klein, P. M. et al. (2021) ‘Detrimental impacts of mixed-ion radiation on nervous system function’, Neurobiology of Disease, 151, p.105252. doi: 10.1016/j.nbd.2021.105252. [PubMed: 33418069]
  103. Klibaite, U. et al. (2017) ‘An unsupervised method for quantifying the behavior of paired animals’, Physical Biology, 14(1), p.015006. doi: 10.1088/1478-3975/aa5c50. [PMC free article: PMC5414632] [PubMed: 28140374]
  104. Krakauer, J. W. et al. (2017) ‘Neuroscience Needs Behavior: Correcting a Reductionist Bias’, Neuron, 93(3), pp. 480–490. doi: 10.1016/j.neuron.2016.12.041. [PubMed: 28182904]
  105. Krizhevsky, A., Sutskever, I. and Hinton, G. E. (2012) ‘ImageNet Classification with Deep Convolutional Neural Networks’,  Advances in Neural Information Processing Systems, 25. Available at: https://proceedings​.neurips​.cc/paper/2012​/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf.
  106. Krook-Magnuson, E. et al. (2013) ‘On-demand optogenetic control of spontaneous seizures in temporal lobe epilepsy’, Nature Communications, 4, pp. 1–8. doi: 10.1038/ncomms2376. [PMC free article: PMC3562457] [PubMed: 23340416]
  107. Kyme, A. et al. (2014) ‘Markerless motion tracking of awake animals in positron emission tomography’, IEEE Transactions on Medical Imaging, 33(11), pp. 2180–2190. doi: 10.1109/TMI.2014.2332821. [PubMed: 24988591]
  108. Lauer, J. et al. (2021) ‘Multi-animal pose estimation and tracking with DeepLabCut’, bioRxiv, pp. 1–20. [PMC free article: PMC9007739] [PubMed: 35414125]
  109. LeCun, Y., Bengio, Y. and Hinton, G. (2015) ‘Deep learning’, Nature, 521(7553), pp. 436–444. doi: 10.1038/nature14539. [PubMed: 26017442]
  110. Leng, X. et al. (2020) ‘Quantifying influence of human choice on the automated detection of Drosophila behavior by a supervised machine learning algorithm’, PLoS ONE, 15(12 December), pp. 1–27. doi: 10.1371/journal.pone.0241696. [PMC free article: PMC7743940] [PubMed: 33326445]
  111. Li, C. et al. (2021) ‘hSMAL: Detailed Horse Shape and Pose Reconstruction for Motion Pattern Recognition’, arXiv preprint arXiv:2106.10102, pp.1–6. Available at: http://arxiv​.org/abs/2106.10102.
  112. Lin, T. Y. et al. (2014) ‘Microsoft COCO: Common objects in context’, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 8693 LNCS(PART 5), pp. 740–755. doi: 10.1007/978-3-319-10602-1_48.
  113. Linderman, S. et al. (2019) ‘Hierarchical recurrent state space models reveal discrete and continuous dynamics of neural activity in C. elegans’, bioRxiv, pp. 1–55. doi: 10.1101/621540.
  114. Liu, M. et al. (2018) ‘Temporal processing and context dependency in caenorhabditis elegans response to mechanosensation’, eLife, 7, pp. 1–29. doi: 10.7554/eLife.36419. [PMC free article: PMC6054533] [PubMed: 29943731]
  115. Liu, X. et al. (2003) ‘Seizure-induced changes in place cell physiology: relationship to spatial memory’, J Neurosci. 2003/12/20, 23(37), pp. 11505–11515. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/14684854. [PMC free article: PMC6740937] [PubMed: 14684854]
  116. Lloyd, S. P. (1982) ‘Least Squares Quantization in PCM’, IEEE Transactions on Information Theory, 28(2), pp. 129–137. doi: 10.1109/TIT.1982.1056489.
  117. Loper, M. et al. (2015) ‘SMPL: A skinned multi-person linear model’, ACM Transactions on Graphics, 34(6), pp. 1–16. doi: 10.1145/2816795.2818013.
  118. van der Maaten, L. and Hinton, G. (2008) ‘Visualizing Data using t-SNE’, Journal of Machine Learning Research, 9(86), pp. 2579–2605. Available at: http://jmlr​.org/papers​/v9/vandermaaten08a.html.
  119. Machado, A. S. et al. (2015) ‘A quantitative framework for whole-body coordination reveals specific deficits in freely walking ataxic mice’, eLife, 4(OCTOBER2015), pp. 1–22. doi: 10.7554/eLife.07892. [PMC free article: PMC4630674] [PubMed: 26433022]
  120. Maghsoudi, O. H. et al. (2018) ‘Application of Superpixels to Segment Several Landmarks in Running Rodents’, Pattern Recognition and Image Analysis, 28(3), pp. 468–482. doi: 10.1134/S1054661818030082.
  121. Markowitz, J. E. et al. (2018) ‘The Striatum Organizes 3D Behavior via Moment-to-Moment Action Selection’, Cell, 174(1), pp. 44–58.e17. doi: 10.1016/j.cell.2018.04.019. [PMC free article: PMC6026065] [PubMed: 29779950]
  122. Marques, J. C. et al. (2018) ‘Structure of the Zebrafish Locomotor Repertoire Revealed with Unsupervised Behavioral Clustering’, Current Biology, 28(2), pp. 181–195.e5. doi: 10.1016/j.cub.2017.12.002. [PubMed: 29307558]
  123. Marshall, J. D. et al. (2021) ‘Continuous Whole-Body 3D Kinematic Recordings across the Rodent Behavioral Repertoire’, Neuron, 109(3), pp. 420–437.e8. doi: 10.1016/j.neuron.2020.11.016. [PMC free article: PMC7864892] [PubMed: 33340448]
  124. Mathis, A. et al. (2018) ‘DeepLabCut: markerless pose estimation of user-defined body parts with deep learning’, Nature Neuroscience, 21(9), pp. 1281–1289. doi: 10.1038/s41593-018-0209-y. [PubMed: 30127430]
  125. Mathis, A. et al. (2020) ‘A Primer on Motion Capture with Deep Learning: Principles, Pitfalls, and Perspectives’, Neuron, 108(1), pp. 44–65. doi: 10.1016/j.neuron.2020.09.017. [PubMed: 33058765]
  126. Mathis, M. W. and Mathis, A. (2020) ‘Deep learning tools for the measurement of animal behavior in neuroscience’, Current Opinion in Neurobiology. 60, pp. 1–11. doi: 10.1016/j.conb.2019.10.008. [PubMed: 31791006]
  127. Matsumoto, J. et al. (2013) ‘A 3D-video-based computerized analysis of social and sexual interactions in rats’, PLoS ONE, 8(10). doi: 10.1371/journal.pone.0078460. [PMC free article: PMC3813688] [PubMed: 24205238]
  128. McInnes, L., Healy, J. and Melville, J. (2018) ‘UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction’. Available at: http://arxiv​.org/abs/1802.03426.
  129. McLachlan, G. and Peel, D. (2000) ‘Finite Mixture Models’, in Wiley Series in Probability and Statistics. doi: 10.1002/0471721182.
  130. Mearns, D. S. et al. (2020) ‘Deconstructing Hunting Behavior Reveals a Tightly Coupled Stimulus-Response Loop’, Current Biology, 30(1), pp. 54–69.e9. doi: 10.1016/j.cub.2019.11.022. [PubMed: 31866365]
  131. Merel, J. et al. (2019) ‘Deep neuroethology of a virtual rodent’, pp. 1–20. Available at: http://arxiv​.org/abs/1911.09451.
  132. Meyer, F. (1994) ‘Topographic distance and watershed lines’, Signal Processing, 38(1), pp. 113–125. doi: 10.1016/0165-1684(94)90060-4.
  133. Mimica, B. et al. (2018) ‘Efficient cortical coding of 3D posture in freely behaving rats’, Science, 362(6414), pp. 584–589. doi: 10.1126/science.aau2013. [PubMed: 30385578]
  134. Mott, M. C., Gordon, J. A. and Koroshetz, W. J. (2018) ‘The NIH BRAIN Initiative: Advancing neurotechnologies, integrating disciplines’, PLoS Biology, 16(11), pp. 1–5. doi: 10.1371/journal.pbio.3000066. [PMC free article: PMC6283590] [PubMed: 30475794]
  135. Nádasdy, Z. et al. (1999) ‘Replay and time compression of recurring spike sequences in the hippocampus’, Journal of Neuroscience, 19(21), pp. 9497–9507. doi: 10.1523/jneurosci.19-21-09497.1999. [PMC free article: PMC6782894] [PubMed: 10531452]
  136. Nath, T. et al. (2019) ‘Using DeepLabCut for 3D markerless pose estimation across species and behaviors’, Nature Protocols, 14(7), pp. 2152–2176. doi: 10.1038/s41596-019-0176-0. [PubMed: 31227823]
  137. Neverova, N. et al. (2021) ‘Discovering Relationships between Object Categories via Universal Canonical Maps’. Available at: http://arxiv​.org/abs/2106.09758.
  138. Newcombe, R. A. et al. (2011) ‘KinectFusion: Real-time dense surface mapping and tracking’, 2011 10th IEEE International Symposium on Mixed and Augmented Reality, ISMAR 2011, pp. 127–136. doi: 10.1109/ISMAR.2011.6092378.
  139. Niessner, M. et al. (2013) ‘Real-Time 3D Reconstruction at Scale Using Voxel Hashing’, ACM Trans. Graph., 32(6), pp.1–11. doi: 10.1145/2508363.2508374.
  140. Nilsson, S. R. O. et al. (2020) ‘Simple Behavioral Analysis (SimBA)—an open source toolkit for computer classification of complex social behaviors in experimental animals’, bioRxiv, 02, pp. 1–29. doi: 10.1101/2020.04.19.049452.
  141. Noebels, J. L. et al. (1990) ‘Stargazer: a new neurological mutant on chromosome 15 in the mouse with prolonged cortical seizures’, Epilepsy Research, 7(2), pp. 129–135. doi: 10.1016/0920-1211(90)90098-G. [PubMed: 2289471]
  142. O’Keefe, J. (1976) ‘Place units in the hippocampus of the freely moving rat’, Experimental Neurology, 51(1), pp. 78–109. doi: 10.1016/0014-4886(76)90055-8. [PubMed: 1261644]
  143. O’Keefe, J. and Dostrovsky, J. (1971) ‘The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat’, Brain Research, 34(1), pp. 171–175. doi: 10.1016/0006-8993(71)90358-1. [PubMed: 5124915]
  144. Olson, E. (2011) ‘AprilTag: A robust and flexible visual fiducial system’, Proceedings—IEEE International Conference on Robotics and Automation, pp. 3400–3407. doi: 10.1109/ICRA.2011.5979561.
  145. Parker, L. et al. (2011) Seizure and epilepsy: Studies of seizure disorders in drosophila, International Review of Neurobiology. 99, p.1–21. doi: 10.1016/B978-0-12-387003-2.00001-X. [PMC free article: PMC3532860] [PubMed: 21906534]
  146. Pasquet, M. O. et al. (2016) ‘Wireless inertial measurement of head kinematics in freely-moving rats’, Scientific Reports, 6(June), pp. 1–13. doi: 10.1038/srep35689. [PMC free article: PMC5073323] [PubMed: 27767085]
  147. Pearre, B. et al. (2017) ‘A fast and accurate zebra finch syllable detector’, PLoS ONE, 12(7), pp. 1–18. doi: 10.1371/journal.pone.0181992. [PMC free article: PMC5533338] [PubMed: 28753628]
  148. Peleh, T. et al. (2019) ‘RFID-supported video tracking for automated analysis of social behaviour in groups of mice’, Journal of Neuroscience Methods, 325(June), p. 108323. doi: 10.1016/j.jneumeth.2019.108323. [PubMed: 31255597]
  149. Pereira, T. D. et al. (2019) ‘Fast animal pose estimation using deep neural networks’, Nature Methods, 16(1), pp. 117–125. doi: 10.1038/s41592-018-0234-5. [PMC free article: PMC6899221] [PubMed: 30573820]
  150. Pereira, T. D. et al. (2020) ‘SLEAP: Multi-animal pose tracking’. bioRxiv, pp.2020–08. doi: https://doi​.org/10.1101/2020.0.
  151. Pereira, T. D., Shaevitz, J. W. and Murthy, M. (2020) ‘Quantifying behavior to understand the brain’, Nature Neuroscience. 23(12), pp. 1537–1549. doi: 10.1038/s41593-020-00734-z. [PMC free article: PMC7780298] [PubMed: 33169033]
  152. Pinel, J. P. and Rovner, L. I. (1978a) ‘Electrode placement and kindling-induced experimental epilepsy’, Exp Neurol. 1978/01/15, 58(2), pp. 335–346. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/618751. [PubMed: 618751]
  153. Pinel, J. P. and Rovner, L. I. (1978b) ‘Experimental epileptogenesis: kindling-induced epilepsy in rats’, Exp Neurol, 58(2), pp. 190–202. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/618743. [PubMed: 618743]
  154. Pisanello, F. et al. (2017) ‘Dynamic illumination of spatially restricted or large brain volumes via a single tapered optical fiber’, Nature Neuroscience, 20(8), pp. 1180–1188. doi: 10.1038/nn.4591. [PMC free article: PMC5533215] [PubMed: 28628101]
  155. Poduri, A. and Whittemore, V. H. (2020) ‘The Benchmarks: Progress and Emerging Priorities in Epilepsy Research’, Epilepsy Currents, 20(1_suppl)20(1_suppl), pp. 3S–4S. doi: 10.1177/1535759719888646. [PMC free article: PMC7031804] [PubMed: 31868039]
  156. Pohl, M. and Mares, P. (1987) ‘Effects of flunarizine on Metrazol-induced seizures in developing rats’, Epilepsy Res. 1987/09/01, 1(5), pp. 302–305. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/3504406. [PubMed: 3504406]
  157. Quigg, M. et al. (1998) ‘Temporal distribution of partial seizures: Comparison of an animal model with human partial epilepsy’, Annals of Neurology, 43(6), pp. 748–755. doi: 10.1002/ana.410430609. [PubMed: 9629844]
  158. Racine, R. J. (1972) ‘Modification of seizure activity by electrical stimulation. II. Motor seizure’, Electroencephalogr Clin Neurophysiol, 32(3), pp. 281–294. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/4110397. [PubMed: 4110397]
  159. Redmon, J. et al. (2016) ‘You only look once: Unified, real-time object detection’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2016-Decem, pp. 779–788. doi: 10.1109/CVPR.2016.91.
  160. Redmon, J. and Farhadi, A. (2018) ‘YOLOv3: An Incremental Improvement’. arXiv preprint arXiv:1804.02767. Available at: http://arxiv​.org/abs/1804.02767.
  161. Ren, X. and Malik, J. (2003) ‘Learning a classification model for segmentation’, Proceedings of the IEEE International Conference on Computer Vision, 1(c), pp. 10–17. doi: 10.1109/iccv.2003.1238308.
  162. Romero-Ferrero, F. et al. (2019) ‘Idtracker.Ai: Tracking All Individuals in Small or Large Collectives of Unmarked Animals’, Nature Methods, 16(2), pp. 179–182. doi: 10.1038/s41592-018-0295-5. [PubMed: 30643215]
  163. Rosenberg, M. et al. (2021) ‘Mice in a labyrinth: Rapid learning, sudden insight, and efficient exploration’, eLife, 10, pp. 1–30. doi: 10.7554/ELIFE.66175. [PMC free article: PMC8294850] [PubMed: 34196271]
  164. Roy, S. et al. (2011) ‘High-precision, three-dimensional tracking of mouse whisker movements with optical motion capture technology’, Frontiers in Behavioral Neuroscience, 5(JUNE), pp. 1–6. doi: 10.3389/fnbeh.2011.00027. [PMC free article: PMC3113147] [PubMed: 21713124]
  165. Rübel, O. et al. (2019) ‘NWB:N 2.0: An Accessible Data Standard for Neurophysiology’, bioRxiv, p. 523035. doi: 10.1101/523035.
  166. Rudolph, S. et al. (2020) ‘Cerebellum-Specific Deletion of the GABAA Receptor δ Subunit Leads to Sex-Specific Disruption of Behavior’, Cell Reports, 33(5), p. 108338. doi: 10.1016/j.celrep.2020.108338. [PMC free article: PMC7700496] [PubMed: 33147470]
  167. Russakovsky, O. et al. (2015) ‘ImageNet Large Scale Visual Recognition Challenge’, International Journal of Computer Vision, 115(3), pp. 211–252. doi: 10.1007/s11263-015-0816-y.
  168. Sainburg, T., Thielk, M. and Gentner, T. Q. (2020) Finding, visualizing, and quantifying latent structure across diverse animal vocal repertoires, PLoS Computational Biology. 16(10), p.e1008228. doi: 10.1371/journal.pcbi.1008228. [PMC free article: PMC7591061] [PubMed: 33057332]
  169. Sanakoyeu, A. et al. (2020) ‘Transferring Dense Pose to Proximal Animal Classes’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 5232–5241. doi: 10.1109/CVPR42600.2020.00528.
  170. Sandbrink, K. J. et al. (2020) ‘Task-driven hierarchical deep neural network models of the proprioceptive pathway’, bioRxiv. doi: 10.1101/2020.05.06.081372.
  171. Sangiamo, D. T., Warren, M. R. and Neunuebel, J. P. (2020) ‘Ultrasonic signals associated with different types of social behavior of mice’, Nature Neuroscience, 23(3), pp. 411–422. doi: 10.1038/s41593-020-0584-z. [PMC free article: PMC7065962] [PubMed: 32066980]
  172. Scheffer, I. E. et al. (2017) ‘ILAE classification of the epilepsies: Position paper of the ILAE Commission for Classification and Terminology’, Epilepsia, 58(4), pp. 512–521. doi: 10.1111/epi.13709. [PMC free article: PMC5386840] [PubMed: 28276062]
  173. Schönberger, J. L. et al. (2016) ‘Pixelwise view selection for unstructured multi-view stereo’, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 9907 LNCS, pp. 501–518. doi: 10.1007/978-3-319-46487-9_31.
  174. Segalin, C. et al. (2020) ‘The Mouse Action Recognition System (MARS): a software pipeline for automated analysis of social behaviors in mice’, bioRxiv. doi: 10.1101/2020.07.26.222299. [PMC free article: PMC8631946] [PubMed: 34846301]
  175. Seitz, S. M. et al. (2006) ‘A comparison and evaluation of multi-view stereo reconstruction algorithms’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1, pp. 519–526. doi: 10.1109/CVPR.2006.19.
  176. Sermanet, P. et al. (2014) ‘Overfeat: Integrated recognition, localization and detection using convolutional networks’, 2nd International Conference on Learning Representations, ICLR 2014—Conference Track Proceedings.
  177. Shuman, T. et al. (2020) ‘Breakdown of spatial coding and interneuron synchronization in epileptic mice’, Nature Neuroscience, 23(2), pp. 229–238. doi: 10.1038/s41593-019-0559-0. [PMC free article: PMC7259114] [PubMed: 31907437]
  178. Skaggs, W. E. and McNaughton, B. L. (1996) ‘Replay of neuronal firing sequences in rat hippocampus during sleep following spatial experience’, Science. 1996/03/29, 271(5257), pp. 1870–1873. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/8596957. [PubMed: 8596957]
  179. Spink, A. J. et al. (2001) ‘The EthoVision video tracking system—A tool for behavioral phenotyping of transgenic mice’, Physiology and Behavior, 73(5), pp. 731–744. doi: 10.1016/S0031-9384(01)00530-3. [PubMed: 11566207]
  180. Stephens, G. J. et al. (2008) ‘Dimensionality and dynamics in the behavior of C. elegans’, PLoS Computational Biology, 4(4), p.e1000028. doi: 10.1371/journal.pcbi.1000028. [PMC free article: PMC2276863] [PubMed: 18389066]
  181. Stephens, G. J. et al. (2011) ‘Emergence of long timescales and stereotyped behaviors in Caenorhabditis elegans’, Proceedings of the National Academy of Sciences of the United States of America, 108(18), pp. 7286–7289. doi: 10.1073/pnas.1007868108. [PMC free article: PMC3088607] [PubMed: 21502536]
  182. Stringer, C. et al. (2019) ‘Spontaneous behaviors drive multidimensional, brainwide activity’, Science, 364(6437), p.eaav7893. doi: 10.1126/science.aav7893. [PMC free article: PMC6525101] [PubMed: 31000656]
  183. Sun, J. J. et al. (2021) ‘The Multi-Agent Behavior Dataset: Mouse Dyadic Social Interactions’, 35th Conference on Neural Information Processing Systems (NeurIPS 2021). Available at: http://arxiv​.org/abs/2104.02710. [PMC free article: PMC11067713] [PubMed: 38706835]
  184. Szigeti, B., Stone, T. and Webb, B. (2016) ‘Inconsistencies in C. elegans behavioural annotation’, bioRxiv. doi: 10.1101/066787.
  185. Taheri, O. et al. (2020) ‘GRAB: A Dataset of Whole-Body Human Grasping of Objects’, in  Vedaldi, A. et al. (eds) Computer Vision—ECCV 2020. Cham: Springer International Publishing, pp. 581–600.
  186. Tao, L. et al. (2019) ‘Statistical structure of locomotion and its modulation by odors’, eLife, 8, pp. 1–30. doi: 10.7554/eLife.41235. [PMC free article: PMC6361587] [PubMed: 30620334]
  187. Teeters, J. L. et al. (2015) ‘Neurodata Without Borders: Creating a Common Data Format for Neurophysiology’, Neuron, 88(4), pp. 629–634. doi: 10.1016/j.neuron.2015.10.025. [PubMed: 26590340]
  188. Tjandrasuwita, M. et al. (2021) ‘Interpreting Expert Annotation Differences in Animal Behavior’. arXiv preprint arXiv:2106.06114. Available at: http://arxiv​.org/abs/2106.06114.
  189. Traynelis, S. F. et al. (2020) ‘Epilepsy Benchmarks Area III: Improved Treatment Options for Controlling Seizures and Epilepsy-Related Conditions Without Side Effects’, Epilepsy Currents, pp. 23S–30S. doi: 10.1177/1535759719895279. [PMC free article: PMC7031805] [PubMed: 31965829]
  190. Triggs, Bill, Philip F. McLauchlan, Richard I. Hartley, and Andrew W. Fitzgibbon. “Bundle adjustment—a modern synthesis.” In Vision Algorithms: Theory and Practice: International Workshop on Vision Algorithms Corfu, Greece, September 21–22, 1999 Proceedings, pp. 298–372. Springer Berlin Heidelberg, 2000.
  191. Tsai, P. T. et al. (2012) ‘Autistic-like behaviour and cerebellar dysfunction in Purkinje cell Tsc1 mutant mice’, Nature, pp. 4–9. doi: 10.1038/nature11310. [PMC free article: PMC3615424] [PubMed: 22763451]
  192. Vanzella, W. et al. (2019) ‘A passive, camera-based head-tracking system for real-time, three-dimensional estimation of head position and orientation in rodents’, Journal of Neurophysiology, 122(6), pp. 2220–2242. doi: 10.1152/jn.00301.2019. [PMC free article: PMC6966308] [PubMed: 31553687]
  193. Vázquez-Guardado, A. et al. (2020) ‘Recent advances in neurotechnologies with broad potential for neuroscience research’, Nature Neuroscience, 23(12), pp. 1522–1536. doi: 10.1038/s41593-020-00739-8. [PubMed: 33199897]
  194. Veliskova, J. et al. (1990) ‘Ketamine suppresses both bicuculline- and picrotoxin-induced generalized tonic-clonic seizures during ontogenesis’, Pharmacol Biochem Behav. 1990/12/01, 37(4), pp. 667–674. Available at: https://www​.ncbi.nlm​.nih.gov/pubmed/2093170. [PubMed: 2093170]
  195. Vogel-Ciernia, A. and Wood, M. A. (2014) ‘Examining object location and object recognition memory in mice’, Curr Protoc Neurosci. 2014/10/08, 69, pp. 8.31.1–17. doi: 10.1002/0471142301.ns0831s69. [PMC free article: PMC4219523] [PubMed: 25297693]
  196. Voigts, J., Sakmann, B. and Celike, T. (2008) ‘Unsupervised whisker tracking in unrestrained behaving animals’, Journal of Neurophysiology, 100(1), pp. 504–515. doi: 10.1152/jn.00012.2008. [PubMed: 18463190]
  197. Wang, J. et al. (2021) ‘Deep 3D human pose estimation: A review’, Computer Vision and Image Understanding, 210, p. 103225. doi: 10.1016/j.cviu.2021.103225.
  198. Wang, Q. et al. (2016) ‘The PSI-U1 snRNP interaction regulates male mating behavior in Drosophila’, Proceedings of the National Academy of Sciences of the United States of America, 113(19), pp. 5269–5274. doi: 10.1073/pnas.1600936113. [PMC free article: PMC4868454] [PubMed: 27114556]
  199. Wang, Y. et al. (2021) ‘Birds of a Feather: Capturing Avian Shape Models from Images’. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14739–14749. Available at: http://arxiv​.org/abs/2105.09396.
  200. Werkhoven, Z. et al. (2019) ‘The structure of behavioral variation within a genotype’, bioRxiv. doi: 10.1101/779363. [PMC free article: PMC8526060] [PubMed: 34664550]
  201. Williams, P. A. et al. (2009) ‘Development of spontaneous recurrent seizures after kainate-induced status epilepticus’, J Neurosci, 29(7), pp. 2103–2112. doi: 10.1523/jneurosci.0980-08.2009. [PMC free article: PMC2897752] [PubMed: 19228963]
  202. Wiltschko, A. B. et al. (2015) ‘Mapping Sub-Second Structure in Mouse Behavior’, Neuron, 88(6), pp. 1121–1135. doi: 10.1016/j.neuron.2015.11.031. [PMC free article: PMC4708087] [PubMed: 26687221]
  203. Wiltschko, A. B. et al. (2020) ‘Revealing the structure of pharmacobehavioral space through motion sequencing’, Nature Neuroscience, 23(11), pp. 1433–1443. doi: 10.1038/s41593-020-00706-3. [PMC free article: PMC7606807] [PubMed: 32958923]
  204. Yamins, D. L. K. et al. (2014) ‘Performance-optimized hierarchical models predict neural responses in higher visual cortex’, Proceedings of the National Academy of Sciences of the United States of America, 111(23), pp. 8619–8624. doi: 10.1073/pnas.1403112111. [PMC free article: PMC4060707] [PubMed: 24812127]
  205. Yamins, D. L. K. and DiCarlo, J. J. (2016) ‘Using goal-driven deep learning models to understand sensory cortex’, Nature Neuroscience, 19(3), pp. 356–365. doi: 10.1038/nn.4244. [PubMed: 26906502]
  206. Ye, S. et al. (2021) ‘SuperAnimal: Improving pre-training for Animal Pose Estimation’, in. CV4Animals: Computer Vision for Animal Behavior Tracking and Modeling. Available at: https://www​.cv4animals.com/paper.
  207. Yemini, E. et al. (2013) ‘A database of Caenorhabditis elegans behavioral phenotypes’, Nature Methods, 10(9), pp. 877–879. doi: 10.1038/nmeth.2560. [PMC free article: PMC3962822] [PubMed: 23852451]
  208. Yosinski, J. et al. (2014) ‘How transferable are features in deep neural networks?’, In Proceedings of the 27th International Conference on Neural Information Processing Systems-Volume 2, pp. 3320–3328.
  209. Zhang, L. et al. (2020) ‘Animal pose estimation from video data with a hierarchical von Mises-Fisher-Gaussian model’, In International conference on artificial intelligence and statistics, pp. 2800–2808. PMLR, 2021.
  210. Zhang, Y. and Park, H. S. (2020) ‘Multiview supervision by registration’, Proceedings—2020 IEEE Winter Conference on Applications of Computer Vision, WACV 2020, 6, pp. 409–417. doi: 10.1109/WACV45572.2020.9093591.
  211. Zimmermann, C. et al. (2020) ‘FreiPose: A Deep Learning Framework for Precise Animal Motion Capture in 3D Spaces’, bioRxiv, pp. 1–38. doi: 10.1101/2020.02.27.967620.
  212. Zuffi, S. et al. (2017) ‘3D menagerie: Modeling the 3D shape and pose of animals’, In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6365–6373. 2017. doi: 10.1109/CVPR.2017.586.
  213. Zuffi, S. et al. (2019) ‘Three-D safari: Learning to estimate zebra pose, shape, and texture from images “in the wild” ’, Proceedings of the IEEE International Conference on Computer Vision, 2019-Octob, pp. 5358–5367. doi: 10.1109/ICCV.2019.00546.
  214. Zuffi, S., Kanazawa, A. and Black, M. J. (2018) ‘Lions and Tigers and Bears: Capturing Non-rigid, 3D, Articulated Shape from Images’, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3955–3963. doi: 10.1109/CVPR.2018.00416.
Copyright Notice

This is an open access publication, available online and distributed under the terms of a Creative Commons Attribution-Non Commercial-No Derivatives 4.0 International licence (CC BY-NC-ND 4.0), a copy of which is available at https://creativecommons.org/licenses/by-nc-nd/4.0/. Subject to this license, all rights are reserved.

Bookshelf ID: NBK609902PMID: 39637108DOI: 10.1093/med/9780197549469.003.0060

Views

  • PubReader
  • Print View
  • Cite this Page
  • PDF version of this title (283M)

Related information

  • PMC
    PubMed Central citations
  • PubMed
    Links to PubMed

Similar articles in PubMed

See reviews...See all...

Recent Activity

Your browsing activity is empty.

Activity recording is turned off.

Turn recording back on

See more...