AnimatLab

Source: Wikipedia, the free encyclopedia.
AnimatLab
Developer(s)David W. Cofer
Gennady Cymbalyuk
James Reid
Ying Zhu
William J. Heitler
Donald H. Edwards
Stable release
2.1.5 / October 5, 2016; 7 years ago (2016-10-05)[1]
Written in
Windows
TypeNeuromechanics
Websitewww.animatlab.com

AnimatLab is an

neural networks that control them to produce behaviors. Users can construct neural models of varied level of details, 3D mechanical models of triangle meshes, and use muscles, motors, receptive fields, stretch sensors and other transducers to interface the two systems. Experiments can be run in which various stimuli are applied and data is recorded, making it a useful tool for computational neuroscience
. The software can also be used to model biomimetic robotic systems.

Motivation

The neuromechanical simulation tool facilitates the construction and testing of

biomimetic robotic
systems.

History

The application was initially developed at Georgia State University under NSF grant #0641326.[3] Version 1 of AnimatLab was released in 2010. Work has continued on the application and a second version was released in June 2013.

Functionality

AnimatLab empowers users to craft models with diverse levels of intricacy, facilitated by a range of available model types.

integrate-and-fire models, or Hodgkin–Huxley models
, with the flexibility to incorporate plugins for additional neuron models. Actuation of joints is achieved using Hill-type muscles, motors, or servos, while adapters interface between neurons and actuators to generate forces. Feedback loops between mechanical components, such as joints, body segments, and muscles, are established through adapters to inform the control system. The platform supports the integration of various stimuli, including voltage clamps, current clamps, and velocity clamps for joints, enabling the design of tailored experiments. Furthermore, comprehensive data recording capabilities allow for the collection of data from different system components, presented through graphical visualization or exported as comma-separated values files for seamless analysis, all within an intuitive graphical user interface.

Neural modeling

A variety of

integrate-and-fire models, and various abstracted firing-rate models are available.[4] This is a relevant feature because the purpose of one's model and its complexity decide which features of neural behavior are important to simulate.[5]

Network construction is graphical, with neurons dragged and dropped into a network and synapses drawn between them. When a synapse is drawn, the user specifies what type to use. Both spiking and non-spiking

Hebbian
) learning mechanisms are available, greatly increasing the capability of the nervous systems constructed.

Rigid body modeling

Body segments are modeled as

Vortex
engine. Users can specify separate collision and graphical meshes for a rigid body, greatly reducing simulation time. In addition, material properties and the interaction between materials can be specified, allowing different restitution, coefficient of friction, etc. within the simulation.

Muscle modeling

A Hill-type muscle model modified according to Shadmehr and Wise[6] can be used for actuation. Muscles are controlled by placing a voltage-tension adapter between a motor neuron and a muscle. Muscles also have stiffness and damping properties, as well as length-tension relationships that govern their behavior. Muscles can are placed to act on muscle attachment bodies in the mechanical simulation, which then apply the muscle tension force to the other bodies in the simulation.

Sensory modeling

Adapters may be placed to convert rigid body measurements to neural activity, much like how voltage-tension adapters are used to activate muscles. These may be joint angles or velocities, rigid body forces or accelerations, or behavioral states (e.g. hunger).

In addition to these scalar inputs, contact fields may be specified on rigid bodies, which then provide pressure feedback to the system. This functionality has been used for skin-like sensing [4] and to detect leg loading in walking structures.[7]

Stimulus types

Stimuli can be applied to mechanical and neural objects in simulation for experimentation. These include current and voltage clamps, as well as velocity clamps for joints between rigid bodies.

Graph types

Data can be output in the form of line graphs and two-dimensional surfaces. Line graphs are useful for most data types, including neural and synaptic output, as well as body and muscle dynamics. Surface plots are useful for outputting activation on contact fields. Both of these can be output as

comma separated values files, allowing the user to use other software such as MATLAB or Excel
for quantitative analysis.

Research performed with AnimatLab

Many academic projects have used AnimatLab to build neuromechanical models and explore behavior. These include:

References

  1. ^ "AnimatLab > Download". animatlab.com. Retrieved 2021-03-25.
  2. ^ "AnimatLab.com - Neuromechanical & Biomechanical Simulation". www.animatlab.com. Retrieved 2024-04-01.
  3. ^ "National Science Foundation Awards". 2010-01-28. Retrieved 2023-11-09.
  4. ^
    S2CID 19398166
    .
  5. .
  6. ^ Shadmehr, Reza; Wise, Steven P. (28 Oct 2004), Computational neurobiology of reaching and pointing: a foundation for motor learning, Cambridge, Massachusetts: MIT Press
  7. ^ a b Szczecinski, N. S. Massively distributed neuromorphic control for legged robots modeled after insect stepping. Master's Thesis. Case Western Reserve University, 2013.
  8. ^ Klishko A., Cofer D. W., Edwards D. H., Prilutsky B. Extremely high paw acceleration during paw shake in the cat: a mechanism revealed by computer simulations. AbstrAm Phys Soc Meeting A38.00007; 2008a.
  9. ^ Klishko A., Prilutsky B., Cofer D. W., Cymbalyuk G., Edwards D. H. Interaction of CPG, spinal reflexes and hindlimb properties in cat paw shake: a computer simulation study. Neuroscience Meeting Planner Online, Program No. 375.12. Society for Neuroscience; 2008b.
  10. ^ Cofer, D. W. (2009). Neuromechanical Analysis of the Locust Jump (Ph.D. dissertation). Available from digital archive database. (Article No. 1056)
  11. PMC 2837733
    .
  12. .
  13. ^ Rinehart M. D., Belanger J. H. Biologically realistic limb coordination during multi-legged walking in the absence of central connections between legs. In: Society for Neuroscience Annual Meeting; 2009.

External links