Lab Home | Phone | Search
Center for Nonlinear Studies  Center for Nonlinear Studies
 Home 
 People 
 Current 
 Affiliates 
 Visitors 
 Students 
 Research 
 ICAM-LANL 
 Publications 
 Conferences 
 Workshops 
 Sponsorship 
 Talks 
 Colloquia 
 Colloquia Archive 
 Seminars 
 Postdoc Seminars Archive 
 Quantum Lunch 
 Quantum Lunch Archive 
 CMS Colloquia 
 Q-Mat Seminars 
 Q-Mat Seminars Archive 
 P/T Colloquia 
 Archive 
 Kac Lectures 
 Kac Fellows 
 Dist. Quant. Lecture 
 Ulam Scholar 
 Colloquia 
 
 Jobs 
 Postdocs 
 CNLS Fellowship Application 
 Students 
 Student Program 
 Visitors 
 Description 
 Past Visitors 
 Services 
 General 
 
 History of CNLS 
 
 Maps, Directions 
 CNLS Office 
 T-Division 
 LANL 
 
Tuesday, January 19, 2010
12:00 PM - 1:30 PM
CNLS Conference Room (TA-3, Bldg 1690)

Seminar

Memory at Different Timescales

Lav R. Varshney
Massachusetts Institute of Technology

Memory operates at numerous timescales, ranging from the fast-scale needed to perform sensory processing to the long-scale needed to preserve human history. In this talk, I will describe three modes of memory and discuss their physical properties, biological implications, and fundamental limits.

Firstly at fast timescales, information may be stored in the dynamic states of neural activity. Using the newly assembled whole, self-consistent C. elegans connectome and linear systems theory, the propagation of neuronal activity in response to sensory stimulation is explored and several activity patterns that could serve as substrates of previously described behaviors are determined. Secondly at intermediate timescales, information may be stored in the synaptic strengths between mammalian neocortical neurons. Experimental investigations have revealed that synapses possess interesting and, in some cases, unexpected properties. I present a theoretical framework based on maximizing information storage capacity of neural tissue under resource constraints that accounts for three of these properties: typical central synapses are noisy, the distribution of synaptic weights among central synapses is wide, and synaptic connectivity between neurons is sparse. Finally at long timescales, information may be written as symbols on a rewriteable medium. For scenarios when representing information and modifying the representation to update the stored message are both expensive, I describe malleable codes and discuss the fundamental trade off between compression efficiency and malleability cost, measured with a string edit distance; the trade off is demarcated by the solution to a graph embedding problem.

Host: Luis Bettencourt, T-5