Cannabis Ruderalis

Content deleted Content added
Commenting on submission (AFCH 0.9.1)
OAbot (talk | contribs)
m Open access bot: arxiv, pmc updated in citation with #oabot.
 
(46 intermediate revisions by 17 users not shown)
Line 1: Line 1:
{{Short description|Israeli physicist}}
{{AFC submission|||u=OffekM|ns=2|ts=20230826114822}} <!-- Do not remove this line! -->

{{AFC comment|1="Follow your dreams"???? [[User:Theroadislong|Theroadislong]] ([[User talk:Theroadislong|talk]]) 12:51, 26 August 2023 (UTC)}}

----

{{Short description|Israeli Physicist}}
{{Draft topics|computing}}
{{AfC topic|blp}}

{{Infobox scientist
{{Infobox scientist
| honorific_prefix = Professor
| honorific_prefix = Professor
Line 32: Line 23:
| siglum =
| siglum =
| pronounce =
| pronounce =
| citizenship = [[Israel]] <!-- use only when necessary per [[WP:INFONAT]] -->
| nationality = [[Israelis|Israeli]] <!-- use only when necessary per [[WP:INFONAT]] -->
| fields = {{plainlist|*Theory of neural networks
| fields = {{plainlist|*Theory of neural networks
*Physical random number generators
*Physical random number generators
Line 72: Line 61:
}}
}}


Ido Kanter (born: 21 Nov. 1959) is an Israeli professor of [[physics]] and the head of the Lab for Reverberating Modes in Neural Networks at the Gonda Brain Research Center at [[Bar-Ilan University]]. He specializes in models of disorder magnetic systems, physical random number generators, theory of [[neural networks]], [[deep learning]] and [[synchronization]] among [[neurons]] and [[lasers]], documented in more than 200 publications.[[File:Ido Kanter Experimental setup.jpg|thumb|left|240px]]
'''Ido Kanter''' (born: 21 Nov. 1959) is an Israeli professor of [[physics]] at and the head of the Lab for Reverberating Modes in Neural Networks at the Gonda Brain Research Center at [[Bar-Ilan University]]. He specializes in models of disorder magnetic systems, physical random number generators, theory of [[neural networks]], [[deep learning]] and [[synchronization]] among [[neurons]] and [[lasers]].
[[File:Ido Kanter Experimental setup.jpg|thumb|303px]]


==Early life and education==
<br>
Kanter was born and raised in [[Rehovot]], [[Israel]] and served in the [[Israeli Defense Force]] from 1978 to 1981.<ref name="About me - contributions"/>
His research on the behavior of artificial networks led him to conduct biological experiments. He made the very unusual transition from theoretical physics to experimental neuroscience as a full professor in 2012, following the teaching of Professor [[P. W. Anderson|Phil Anderson]], a [[Nobel prize|Nobel laureate]], his supervisor during his postdoc at [[Princeton University]]: Follow your dreams.


He attended [[Bar-Ilan University]] and graduated with a bachelor's degree in [[physics]] and [[computer science]] in 1983. In 1987, he received his [[Ph.D.]] from Bar-Ilan University. His [[thesis]] was ''Theory of Spin Glasses and its Applications to Complex Problems in Mathematics and Biology,'' under the supervision of Professor [[Haim Sompolinsky]].<ref name="About me - contributions">{{cite web |title=About me |url=https://kanterlabsite.wixsite.com/idokanter/about-me |publisher=Kanter Lab |at=Download Main Contributions |access-date=25 April 2024}}</ref>
==Education==
Ido Kanter was born and raised in [[Rehovot]], [[Israel]] and served in the [[Israeli Defense Force]] from 1978 to 1981. Kanter graduated summa cum laude from [[Bar-Ilan University]] with a bachelor's degree in [[physics]] and [[computer science]] in 1983. In 1987, he received his direct [[Ph.D.]] from Bar-Ilan University with his [[thesis]], “Theory of Spin Glasses and its Applications to Complex Problems in Mathematics and Biology,” under the supervision of Professor [[Haim Sompolinsky]].

==Academic Career==
After completing his Ph.D., Kanter joined Professor [[P. W. Anderson|Phil W. Anderson’s]] group at [[Princeton University]] as a visiting research fellow (1988-1989). He was also a visiting research fellow at [[AT&T Bell Labs]], collaborating with [[Yann LeCun]] (1989).
In 1989, Kanter became a senior lecturer at Bar-Ilan University. He became an associate professor in 1991 at age 31 and a full professor in 1996.


He was a visiting research fellow at Princeton University from 1988 to 1989, working with [[Phil W. Anderson]]. He was also a visiting research fellow at AT&T Bell Labs, with [[Yann LeCun|Yann le Cun]], then 1989 joined the physics department at Bar-Ilan University in 1989.<ref name="About me - contributions"/>
==Research==
==Research==
Ido Kanter specializes in models of disorder magnetic systems, ultrafast physical random number generators, theory of neural networks, neural cryptography, deep learning and synchronization among neurons and lasers and experimental and theoretical neuroscience, documented in more than 220 publications.<ref>[https://scholar.google.com/citations?user=0MdAUb0AAAAJ&hl=en Ido Kanter's Google Scholar profile]</ref>
The main discoveries and achievements of Kanter's research include:


==Main contributions==
'''1. Ultra-fast Physical Random Number Generators'''
[[File:New brain learning.webm|thumb|Dendritic learning as an alternative to synaptic plasticity (with audio)|thumbtime=0|289x289px]]Using a combination of theoretical and experimental methods,<ref>{{cite web | url=https://kanterlabsite.wixsite.com/idokanter/about-me | title=About Me }}</ref> Kanter has made contributions to various fields ranging from statistical physics and communication to neural cryptography and neuroscience.<ref>{{cite web | url=https://physics.biu.ac.il/en/node/578 | title=Kanter Ido &#124; Department of Physics }}</ref> These include work on a field of statistical physics known as the inverse problem,<ref>{{cite journal | last=Kanter | first=I. | last2=Gotesdyner | first2=R. | title=Do classical spin systems with the same metastable states have identical Hamiltonians? | journal=Physical Review Letters | volume=72 | issue=17 | date=1994 | doi=10.1103/PhysRevLett.72.2678 | pages=2678–2681}}</ref> bridging between Shannon theory and the second thermodynamic law,<ref>{{cite journal | last=Shental | first=O. | last2=Kanter | first2=I. | title=Shannon meets Carnot: Generalized second thermodynamic law | journal=EPL (Europhysics Letters) | volume=85 | issue=1 | date=2009 | doi=10.1209/0295-5075/85/10006 | page=10006| arxiv=0806.3763 }}</ref> presenting a cryptographic key exchange protocol based on neural networks,<ref>{{cite journal | last=Kanter | first=Ido | last2=Kopelowitz | first2=Evi | last3=Kinzel | first3=Wolfgang | title=Public Channel Cryptography: Chaos Synchronization and Hilbert’s Tenth Problem | journal=Physical Review Letters | volume=101 | issue=8 | date=2008 | doi=10.1103/PhysRevLett.101.084102| arxiv=0806.0931 }}</ref> and creating an ultrafast non-deterministic random bit generator (RBG).<ref>{{cite journal | last1=Kanter | first1=Ido | last2=Aviad | first2=Yaara | last3=Reidler | first3=Igor | last4=Cohen | first4=Elad | last5=Rosenbluh | first5=Michael | title=An optical ultrafast random bit generator | journal=Nature Photonics | volume=4 | issue=1 | date=2010 | doi=10.1038/nphoton.2009.235 | pages=58–61| bibcode=2010NaPho...4...58K }}</ref>
[[File:Random number generator.jpg|thumb|Scheme of a chaotic laser (Top) and a random number generator (Bottom).]]
Random bit sequences are a necessary element in many aspects of our life including [[secure communication]], [[Monte Carlo simulations]], [[stochastic modeling]], and [[on-line-gaming]]. For decades, they were generated at high data rates using deterministic algorithms based on a secret seed, called pseudorandom bit generators. However, their unpredictability is limited by their deterministic origin. [[Nondeterministic random bit generators (RBGs)]] rely on stochastic physical processes, though their generation rates were below 100 Mbit/s, a few orders of magnitude below that of the deterministic algorithms, until 2008. The feasibility of fast stochastic physical RBGs was an open question for decades. A breakthrough occurred in late 2008 when Professor Atsushi Uchida combined binary digitization of two independent chaotic semiconductor lasers (SLs), achieving 1.7 Gbit/s RBG. Months later, a 12.5 Gbit/s RBG based on a single chaotic SL was achieved by Kanter et al..<ref>Reidler, I., Aviad, Y., Rosenbluh, M. & Kanter, I. [https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.103.024102 Ultrahigh-speed random number generation based on a chaotic semiconductor laser.] Physical review letters 103, 024102 (2009). doi.org/10.1103/PhysRevLett.103.024102</ref>, and in late 2009, its generation rate was increased to 300 Gbit/s<ref>Kanter, I., Aviad, Y., Reidler, I., Cohen, E. & Rosenbluh, M. [https://www.nature.com/articles/nphoton.2009.235 An optical ultrafast random bit generator.] Nature Photonics 4, 58-61 (2010).
doi.org/10.1038/nphoton.2009.235</ref>. This method was robust to perturbations and control parameters and is used in many applications, including on chips.


Kanter is currently focusing on the field of experimental and theoretical neuroscience, Kanter studies a variety of topics including the new neuron,<ref>{{cite journal | last=Sardi | first=Shira | last2=Vardi | first2=Roni | last3=Sheinin | first3=Anton | last4=Goldental | first4=Amir | last5=Kanter | first5=Ido | title=New Types of Experiments Reveal that a Neuron Functions as Multiple Independent Threshold Units | journal=Scientific Reports | publisher=Springer Science and Business Media LLC | volume=7 | issue=1 | date=2017 | doi=10.1038/s41598-017-18363-1| pmc=5740076 }}</ref> dendritic learning,<ref>{{cite journal | last=Sardi | first=Shira | last2=Vardi | first2=Roni | last3=Goldental | first3=Amir | last4=Tugendhaft | first4=Yael | last5=Uzan | first5=Herut | last6=Kanter | first6=Ido | title=Dendritic Learning as a Paradigm Shift in Brain Learning | journal=ACS Chemical Neuroscience | volume=9 | issue=6 | date=2018 | doi=10.1021/acschemneuro.8b00204 | pages=1230–1232}}</ref> neural interfaces, and machine learning.<ref>{{cite web | url=https://gondabrain.biu.ac.il/en/node/317 | title=Reverberating Modes in Neural Networks &#124; the Gonda Multidisciplinary Brain Research Center }}</ref>
'''2. The New Neuron'''
[[File:New brain learning.webm|thumb|Dendritic learning as an alternative to synaptic plasticity (with audio)|thumbtime=0]]
Neurons are the basic computational building blocks that compose the brain. According to the neuronal computational scheme, which has been used since the beginning of the 20<sup>th</sup> century, each neuron functions as a centralized excitable element. The neuron accumulates its incoming electrical signals from connecting neurons through several terminals (dendrites) and generates a short electrical pulse, known as a spike, when its threshold is crossed.
Using new types of experiments on neuronal cultures, Kanter and his experimental research group have demonstrated that this century-old assumption regarding brain activity is mistaken. They showed that each neuron functions as a collection of excitable elements<ref name="3.">Sardi, S., Vardi, R., Sheinin, A., Goldental, A. & Kanter, I. [https://www.nature.com/articles/s41598-017-18363-1 New types of experiments reveal that a neuron functions as multiple independent threshold units.] Sci Rep-Uk 7, 18036 (2017). doi.org/10.1038/s41598-017-18363-1</ref>, where each excitable element is sensitive to the directionality of the origin of the input signal. Two weak inputs from different directions (different dendrites) will not sum up to generate a spike, while two weak inputs from the same direction (same dendrite) will generate a spike.
This neuronal anisotropic feature was experimentally extended to the following anisotropic neuronal reversible plasticity: neuronal response latency<ref>Vardi, R., Timor, R., Marom, S., Abeles, M. & Kanter, I. [https://iopscience.iop.org/article/10.1209/0295-5075/100/48003 Synchronization with mismatched synaptic delays: A unique role of elastic neuronal latency.] EPL (Europhysics Letters) 100, 48003 (2012). doi.org/10.1209/0295-5075/100/48003</ref>, response failures<ref>Vardi, R., Goldental, A., Marmari, H., Brama, H., Stern, E. A., Sardi, S., Sabo, P. & Kanter, I. [https://www.frontiersin.org/articles/10.3389/fncir.2015.00029/full Neuronal response impedance mechanism implementing cooperative networks with low firing rates and μs precision.] Frontiers in neural circuits 9, 29 (2015).
doi.org/10.3389/fncir.2015.00029</ref>, absolute refractory periods<ref>Sardi, S., Vardi, R., Tugendhaft, Y., Sheinin, A., Goldental, A. & Kanter, I. [https://journals.aps.org/pre/abstract/10.1103/PhysRevE.105.014401 Long anisotropic absolute refractory periods with rapid rise times to reliable responsiveness.] Physical Review E 105, 014401 (2022). doi.org/10.1103/PhysRevE.105.014401</ref><ref>Vardi, R., Tugendhaft, Y., Sardi, S. & Kanter, I. [https://iopscience.iop.org/article/10.1209/0295-5075/ac177a Significant anisotropic neuronal refractory period plasticity.] EPL (Europhysics Letters) 134, 60007 (2021).
doi.org/10.1209/0295-5075/ac177a</ref>, and spike waveforms.<ref name="3." /> The timescale of reversible plasticity is broadband, ranging between microseconds and tens of seconds.


==Selected publications==
'''3. Dendritic Learning as a Paradigm Shift in Brain Learning'''
* {{cite journal | last=Gross | first=D. J. | last2=Kanter | first2=I. | last3=Sompolinsky | first3=H. | title=Mean-field theory of the Potts glass | journal=Physical Review Letters | volume=55 | issue=3 | date=1985 | doi=10.1103/PhysRevLett.55.304 | pages=304–307}}
* {{cite journal | last1=Sompolinsky | first1=H. | last2=Kanter | first2=I. | title=Temporal Association in Asymmetric Neural Networks | journal=Physical Review Letters | volume=57 | issue=22 | date=1986| doi=10.1103/PhysRevLett.57.2861 | pages=2861–2864| pmid=10033885 | bibcode=1986PhRvL..57.2861S }}
* {{cite journal | last1=Kanter | first1=I. | last2=Sompolinsky | first2=H. | title=Associative recall of memory without errors | journal=Physical Review A | volume=35 | issue=1 | date=1987| doi=10.1103/PhysRevA.35.380 | pages=380–392| pmid=9897963 | bibcode=1987PhRvA..35..380K }}
* {{cite journal | last=Kanter | first=Ido | title=Potts-glass models of neural networks | journal=Physical Review A | volume=37 | issue=7 | date=1988 | doi=10.1103/PhysRevA.37.2739 | pages=2739–2742}}
* {{cite journal | last=Kanter | first=I | last2=Kinzel | first2=W | last3=Kanter | first3=E | title=Secure exchange of information by synchronization of neural networks | journal=Europhysics Letters (EPL) | volume=57 | issue=1 | date=2002 | doi=10.1209/epl/i2002-00552-9 | pages=141–147| arxiv=cond-mat/0202112 }}
* {{cite journal | last1=Reidler | first1=I. | last2=Aviad | first2=Y. | last3=Rosenbluh | first3=M. | last4=Kanter | first4=I. | title=Ultrahigh-Speed Random Number Generation Based on a Chaotic Semiconductor Laser | journal=Physical Review Letters | volume=103 | issue=2 | date=2009 | page=024102 | doi=10.1103/PhysRevLett.103.024102| pmid=19659208 | bibcode=2009PhRvL.103b4102R }}
* {{cite journal | last1=Kanter | first1=Ido | last2=Aviad | first2=Yaara | last3=Reidler | first3=Igor | last4=Cohen | first4=Elad | last5=Rosenbluh | first5=Michael | title=An optical ultrafast random bit generator | journal=Nature Photonics | volume=4 | issue=1 | date=2010 | doi=10.1038/nphoton.2009.235 | pages=58–61| bibcode=2010NaPho...4...58K }}


==References==
In 1949, Donald Hebb's pioneering work suggested that learning occurs in the brain by modifying the strength of the synapses, whereas neurons function as the computational elements in the brain. This has remained the common assumption in neuroscience and initiated the research field of machine learning decades ago.
{{Reflist}}
Using new experiments focused on neuronal cultures, Kanter’s research group revealed a new underlying mechanism for fast (several seconds) brain learning process, dendritic learning<ref>Sardi, S., Vardi, R., Meir, Y., Tugendhaft, Y., Hodassman, S., Goldental, A. & Kanter, I. [https://www.nature.com/articles/s41598-020-63755-5 Brain experiments imply adaptation mechanisms which outperform common AI learning algorithms.] Sci Rep-Uk 10, 1-10 (2020) doi.org/10.1038/s41598-020-63755-5.</ref><ref>Sardi, S., Vardi, R., Goldental, A., Sheinin, A., Uzan, H. & Kanter, I. [https://www.nature.com/articles/s41598-018-23471-7 Adaptive nodes enrich nonlinear cooperative learning beyond traditional adaptation by links.] Sci Rep-Uk 8, 5100, doi:10.1038/s41598-018-23471-7 (2018). doi.org/10.1038/s41598-018-23471-7</ref><ref>Sardi, S., Vardi, R., Goldental, A., Tugendhaft, Y., Uzan, H. & Kanter, I. [https://pubs.acs.org/doi/10.1021/acschemneuro.8b00204 Dendritic learning as a paradigm shift in brain learning.] ACS chemical neuroscience 9, 1230-1232 (2018). doi.org/10.1021/acschemneuro.8b00204</ref>. This fast mechanism opposed the previous common belief based solely on slow (tens of minutes) synaptic plasticity. The presented paradigm indicates that fast dendritic learning occurs in closer proximity to the neuron, the computational unit. It presents a new type of abundant cooperative nonlinear dynamics, as effectively all incoming synapses to an updated dendrite concurrently undergo the same adaptation. In addition, dendritic strengths are self-oscillating<ref>Uzan, H., Sardi, S., Goldental, A., Vardi, R. & Kanter, I. [https://www.nature.com/articles/s41598-018-31523-1 Stationary log-normal distribution of weights stems from spontaneous ordering in adaptive node networks.] Sci Rep-Uk 8, 13091 (2018).
doi.org/10.1038/s41598-018-31523-1</ref>, and weak synapses, which comprise the majority of our brain and were previously assumed to be insignificant, may play a key role in brain activity.


==External links==
'''4. The Inverse Problem'''
* [https://physics.biu.ac.il/en/node/578 Faculty website]
[[File:Inverse problem picture.png|thumb|Inverse problem scheme (blue).]]
* [https://kanterlabsite.wixsite.com/idokanter Ido Kanter Lab]
The usual procedure of statistical physics is aimed at the calculation of macroscopic quantities (such as pressure and magnetization) on the basis of model parameters (Hamiltonian). In the inverse problem this procedure is reversed; one seeks for the existence of parameters of a model based on the required statistical properties. Kanter was a pioneer of this field of research. In 1994, he asked the question, "Do classical spin systems with the same metastable states have identical Hamiltonians?"<ref>Kanter, I. & Gotesdyner, R. [https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.72.2678 Do classical spin systems with the same metastable states have identical Hamiltonians?] Physical review letters 72, 2678 (1994). doi.org/10.1103/PhysRevLett.72.2678</ref>. A year later he generalized this method to neural networks<ref>Kanter, I., Kessler, D., Priel, A. & Eisenstein, E. [https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.75.2614 Analytical study of time series generation by feed-forward networks.] Physical review letters 75, 2614 (1995).
doi.org/10.1103/PhysRevLett.75.2614</ref> and then to the Simplex method<ref>Keren, S., Kfir, H. & Kanter, I. [https://iopscience.iop.org/article/10.1088/0305-4470/39/16/004 Possible sets of autocorrelations and the simplex algorithm.] Journal of Physics A: Mathematical and General 39, 4161 (2006).
doi.org/10.1088/0305-4470/39/16/004</ref>. These types of inverse problems opened a new research horizon in statistical mechanics, followed by numerous physical system analyses and interdisciplinary directions.


{{DEFAULTSORT:Kanter, Ido}}
'''5. Shannon Meets Carnot: Generalized Second Thermodynamic Law'''[[File:Shannon meets Carnot.jpg|thumb|Top: A temperature-dependent spring constant at High/Cold (T<sub>H</sub>/T<sub>C</sub>) temperatures. Bottom: Carnot cycle (black), information heat-engine (red).|150px]]
[[Category:1959 births]]
The laws of thermodynamics describe the transport of heat and work in macroscopic processes and play a fundamental role in the physical sciences. In particular, the second thermodynamic law linearly relates the change in the entropy, dS, to the amount of heat, dQ, absorbed by a system at equilibrium, dQ = TdS, thus defining the temperature, T, of the system.
[[Category:Living people]]
The prior scientific belief was that information theory is primarily a mathematical creature and has its own vitality independent of the physical laws of nature. Kanter, together with his Ph.D student Ori Shental, were pioneers in bridging between Shannon theory and the second law of thermodynamics<ref>Shental, O. & Kanter, I. [https://iopscience.iop.org/article/10.1209/0295-5075/85/10006/meta Shannon meets Carnot: Generalized second thermodynamic law.] Europhysics Letters 85, 10006 (2009).
doi.org/10.1209/0295-5075/85/10006</ref> and in fusing together statistical physics and information and communication theories<ref name="16.">Peleg, Y., Efraim, H., Shental, O. & Kanter, I. [https://iopscience.iop.org/article/10.1088/1742-5468/2010/01/P01014 Mutual information via thermodynamics: three different approaches.] Journal of Statistical Mechanics: Theory and Experiment 2010, P01014 (2010).
doi.org/10.1088/1742-5468/2010/01/P01014</ref>
In this line of work, Shental and Kanter had generalized the second thermodynamic law to encompass systems with temperature-dependent Hamiltonians and obtained the generalized thermodynamic law <math>dQ = TdS + <\frac{dE}{dT}>dT</math>, where < · > denotes averaging over the Boltzmann distribution. <br>
This generalized second law of thermodynamics reveals a new definition to the basic notion of temperature and suggests an information heat engine<ref name="16." />. Interestingly, it provides a quantitative bridge between the realm of thermodynamics and information theory in the context of communication channels, such as the popular Gaussian and binary symmetric channels. Therefore, purely information-theoretic measures such as entropy, mutual information, and channel capacity, are correctly re-derived from thermodynamics.

'''6. Zero-lag Synchronization and the Greatest Common Divisor of Network Loops'''
[[File:Zero-lag Synchronization and the Greatest Common Divisor of Network Loops.png|thumb|Nodes with the same color are in zero-lag synchronization, GCD=1 (Top) GCD=5 (Bottom)|150px]]
The emergence of zero-lag synchronization among distant excitable or chaotic units, without a common input, remained a puzzle for many years. Between 2008-2010, a new non-local mechanism for zero-lag synchronization of a network composed of chaotic units with time-delay couplings was presented by Kanter and Professor Wolfgang Kinzel. The non-local mechanism is the greatest common divisor (GCD) of network's loops<ref>Kanter, I., Kopelowitz, E., Vardi, R., Zigzag, M., Kinzel, W., Abeles, M. & Cohen, D. [https://iopscience.iop.org/article/10.1209/0295-5075/93/66001 Nonlocal mechanism for cluster synchronization in neural circuits.] Europhysics Letters 93, 66001 (2011). doi.org/10.1209/0295-5075/93/66001</ref>. For GCD=1 of network's loops, all units are in zero-lag; for GCD>1, the network splits into GCD-clusters in which clustered units are in zero-lag synchronization. These results are supported by simulations of chaotic networks, mixing arguments, and analytical solutions.
The zero-lag synchronization of distant systems and the non-local GCD mechanism were experimentally tested on the reverberating activity patterns embedded in networks of cortical neurons.<ref>Vardi, R., Wallach, A., Kopelowitz, E., Abeles, M., Marom, S. & Kanter, I. [https://iopscience.iop.org/article/10.1209/0295-5075/97/66002 Synthetic reverberating activity patterns embedded in networks of cortical neurons.] Europhysics Letters 97, 66002 (2012). doi.org/10.1209/0295-5075/97/66002</ref><ref>Vardi, R., Timor, R., Marom, S., Abeles, M. & Kanter, I. [https://iopscience.iop.org/article/10.1209/0295-5075/100/48003 Synchronization with mismatched synaptic delays: A unique role of elastic neuronal latency.] Europhysics Letters 100, 48003 (2012). doi.org/10.1209/0295-5075/100/48003</ref>, controlling synchronization in large laser networks<ref>Nixon, M., Fridman, M., Ronen, E., Friesem, A. A., Davidson, N. & Kanter, I. [https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.108.214101 Controlling synchronization in large laser networks.] Physical review letters 108, 214101 (2012).
doi.org/10.1103/PhysRevLett.108.214101</ref><ref>Nixon, M., Friedman, M., Ronen, E., Friesem, A. A., Davidson, N. & Kanter, I. [https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.106.223901 Synchronized cluster formation in coupled laser networks.] Physical review letters 106, 223901 (2011). doi.org/10.1103/PhysRevLett.106.223901</ref>, synchronization in small networks of time-delay coupled chaotic diode lasers<ref>Aviad, Y., Reidler, I., Zigzag, M., Rosenbluh, M. & Kanter, I. [https://opg.optica.org/oe/fulltext.cfm?uri=oe-20-4-4352&amp;id=227495 Synchronization in small networks of time-delay coupled chaotic diode lasers.] Opt Express 20, 4352-4359 (2012).
doi.org/10.1364/OE.20.004352</ref>, and in chaos synchronization in networks of semiconductor superlattices<ref>Li, W., Aviad, Y., Reidler, I., Song, H., Huang, Y., Biermann, K., Rosenbluh, M., Zhang, Y., Grahn, H. T. & Kanter, I. [https://iopscience.iop.org/article/10.1209/0295-5075/112/30007 Chaos synchronization in networks of semiconductor superlattices.] Europhysics Letters 112, 30007 (2015). doi.org/10.1209/0295-5075/112/30007</ref>

'''7. Neural Cryptography'''
[[File:Neural Cryptography.jpg|thumb|Synchronized weights (key) between two parity machines by mutual communication. ]]
A bridge between the synchronization of mutual coupling of two feedforward networks<ref>Kinzel, W., Metzler, R. & Kanter, I. [https://iopscience.iop.org/article/10.1088/0305-4470/33/14/101 Dynamics of interacting neural networks.] Journal of Physics A: Mathematical and General 33, L141 (2000). doi.org/10.1088/0305-4470/33/14/101</ref> and the generation of symmetric key-exchange protocol over a public channel was established by Kanter and Professor Wolfgang Kinzel in 2001<ref>Kinzel, W. & Kanter, I. [https://link.springer.com/chapter/10.1007/3-540-45618-x_30 Interacting neural networks and cryptography, in Advances in solid state physics] 383-391 (Springer, 2002). doi.org/10.1007/3-540-45618-X_30</ref>. A passive attacker who knows the protocol and all details of any transmission of the data will find it difficult to reveal the mutual key. Simulations and analytical work indicate that synchronization is superior to tracking by a passive attacker.
This type of key-exchange protocol was extended to the synchronization process of two mutually delayed coupled deterministic chaotic maps<ref name="26.">Kanter, I., Kopelowitz, E. & Kinzel, W. [https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.101.084102 Public channel cryptography: chaos synchronization and Hilbert’s tenth problem.] Phys Rev Lett 101, 084102 (2008). doi.org/10.1103/PhysRevLett.101.084102</ref><ref>Mislovaty, R., Klein, E., Kanter, I. & Kinzel, W. [https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.91.118701 Public channel cryptography by synchronization of neural networks and chaotic maps.] Physical review letters 91, 118701 (2003). doi.org/10.1103/PhysRevLett.91.118701</ref> and was examined experimentally using two mutual chaotic semiconductor lasers<ref>Klein, E., Gross, N., Rosenbluh, M., Kinzel, W., Khaykovich, L. & Kanter, I. [https://journals.aps.org/pre/abstract/10.1103/PhysRevE.73.066214 Stable isochronal synchronization of mutually coupled chaotic lasers.] Phys Rev E 73, 066214 (2006).
doi.org/10.1103/PhysRevE.73.066214</ref><ref>Kanter, I., Gross, N., Klein, E., Kopelowitz, E., Yoskovits, P., Khaykovich, L., Kinzel, W. & Rosenbluh, M.
[https://journals.aps.org/prl/abstract/10.1103/PhysRevLett.98.154101 Synchronization of mutually coupled chaotic lasers in the presence of a shutter.] Physical review letters 98, 154101 (2007).
doi.org/10.1103/PhysRevLett.98.154101</ref><ref>Kanter, I., Butkovski, M., Peleg, Y., Zigzag, M., Aviad, Y., Reidler, I., Rosenbluh, M. & Kinzel, W. [https://opg.optica.org/aop/fulltext.cfm?uri=oe-18-17-18292&amp;id=205049 Synchronization of random bit generators based on coupled chaotic lasers and application to cryptography.] Opt Express 18, 18292-18302 (2010).
doi.org/10.1364/OE.18.018292</ref>.
The task of a passive attacker is mapped in some limits onto Hilbert’s tenth problem<ref name="26." />, solving a set of nonlinear Diophantine equations, which was proven to be in the class of NP-complete problems. This bridge between nonlinear dynamics and NP-complete problems opens a horizon for new types of secure public-channel protocols.

'''8. Physics Assists with Key Challenges in Artificial Intelligence'''
[[File:Brain AI.webm|thumb|The brain is a source for a new AI mechanism (with audio)|thumbtime=0]]
From 2020-2023, Kanter's research shed light on the following fundamental theoretical questions regarding deep learning (DL):

(a) Is shallow learning equivalent to DL? Is DL a necessary ingredient for artificial intelligence (AI)?<ref>Meir, Y., Tevet, O., Tzach, Y., Hodassman, S., Gross, R. D. & Kanter, I. [https://www.nature.com/articles/s41598-023-32559-8 Efficient shallow learning as an alternative to deep learning.] Scientific Reports 13, 5423 (2023).
doi.org/10.1038/s41598-023-32559-8</ref>

(b) What is the best location of the pooling operators to enhance accuracy?<ref>Meir, Y., Tzach, Y., Gross, R. D., Tevet, O., Vardi, R. & Kanter, I. [https://arxiv.org/abs/2303.05800 Enhancing the success rates by performing pooling decisions adjacent to the output layer.] Scientific Reports </ref>

(c) Is brain learning, based on tree architectures only, weaker than AI?<ref name="33.">Meir, Y., Ben-Noam, I., Tzach, Y., Hodassman, S. & Kanter, I. [https://www.nature.com/articles/s41598-023-27986-6 Learning on tree architectures outperforms a convolutional feedforward network.] Scientific Reports 13, 962 (2023).
doi.org/10.1038/s41598-023-27986-6</ref>

(d) Is there a universal law regarding how to efficiently build DL architectures?<ref>Meir, Y., Sardi, S., Hodassman, S., Kisos, K., Ben-Noam, I., Goldental, A. & Kanter, I. [https://www.nature.com/articles/s41598-020-76764-1 Power-law scaling to assist with key challenges in artificial intelligence.] Scientific reports 10, 19628 (2020).
doi.org/10.1038/s41598-020-76764-1</ref>

(e) Can error rates follow a universal law as a function of dataset sizes?<ref name="33." />

(f) What is the mechanism underlying DL?<ref>Tzach, Y., Meir, Y., Tevet, O., Gross, R. D., Hodassman, S., Vardi, R. & Kanter, I. [https://arxiv.org/abs/2305.18078 The mechanism underlying successful deep learning.] arXiv preprint arXiv:2305.18078 (2023). </ref>

==External Links==
1. Ido kanter's personal website: https://kanterlabsite.wixsite.com/idokanter

2. Ido Kanter's Bar-Ilan website: https://physics.biu.ac.il/en/node/578

3. Selected press releases: https://kanterlabsite.wixsite.com/idokanter/press-articles

4. Movie on “Dendritic learning as an alternative to synaptic plasticity”: https://vimeo.com/702894966

5. Movie on “Unreliable neurons improve brain functionalities and cryptography”: https://vimeo.com/752532666

6. Ido Kanter's Google scholar: https://scholar.google.com/citations?user=0MdAUb0AAAAJ&hl=en

7. Ido Kanter's ResearchGate: https://www.researchgate.net/profile/Ido-Kanter

8. Ido Kanter's Linkdin: https://www.linkedin.com/in/ido-kanter-8448a016/recent-activity/all/

==References==
{{Reflist|30em}}

Latest revision as of 06:48, 29 April 2024

Professor
Ido Kanter
Professor Ido Kanter
Born (1959-11-21) November 21, 1959 (age 64)
Alma materBar-Ilan University
Awards
Weizmann Postdoctoral Fellowship (1988-1989)

Humboldt Senior Research Prize (2001)

Scientific career
Fields
  • Theory of neural networks
  • Physical random number generators
  • Neuroscience in-vitro
  • Deep learning
  • Synchronization of neurons and lasers
  • Neural cryptography
InstitutionsPostdoc: Princeton University, with P. W. Anderson
Doctoral advisorHaim Sompolinsky

Ido Kanter (born: 21 Nov. 1959) is an Israeli professor of physics at and the head of the Lab for Reverberating Modes in Neural Networks at the Gonda Brain Research Center at Bar-Ilan University. He specializes in models of disorder magnetic systems, physical random number generators, theory of neural networks, deep learning and synchronization among neurons and lasers.

Early life and education[edit]

Kanter was born and raised in Rehovot, Israel and served in the Israeli Defense Force from 1978 to 1981.[1]

He attended Bar-Ilan University and graduated with a bachelor's degree in physics and computer science in 1983. In 1987, he received his Ph.D. from Bar-Ilan University. His thesis was Theory of Spin Glasses and its Applications to Complex Problems in Mathematics and Biology, under the supervision of Professor Haim Sompolinsky.[1]

He was a visiting research fellow at Princeton University from 1988 to 1989, working with Phil W. Anderson. He was also a visiting research fellow at AT&T Bell Labs, with Yann le Cun, then 1989 joined the physics department at Bar-Ilan University in 1989.[1]

Research[edit]

Ido Kanter specializes in models of disorder magnetic systems, ultrafast physical random number generators, theory of neural networks, neural cryptography, deep learning and synchronization among neurons and lasers and experimental and theoretical neuroscience, documented in more than 220 publications.[2]

Main contributions[edit]

Dendritic learning as an alternative to synaptic plasticity (with audio)

Using a combination of theoretical and experimental methods,[3] Kanter has made contributions to various fields ranging from statistical physics and communication to neural cryptography and neuroscience.[4] These include work on a field of statistical physics known as the inverse problem,[5] bridging between Shannon theory and the second thermodynamic law,[6] presenting a cryptographic key exchange protocol based on neural networks,[7] and creating an ultrafast non-deterministic random bit generator (RBG).[8]

Kanter is currently focusing on the field of experimental and theoretical neuroscience, Kanter studies a variety of topics including the new neuron,[9] dendritic learning,[10] neural interfaces, and machine learning.[11]

Selected publications[edit]

References[edit]

  1. ^ a b c "About me". Kanter Lab. Download Main Contributions. Retrieved 25 April 2024.
  2. ^ Ido Kanter's Google Scholar profile
  3. ^ "About Me".
  4. ^ "Kanter Ido | Department of Physics".
  5. ^ Kanter, I.; Gotesdyner, R. (1994). "Do classical spin systems with the same metastable states have identical Hamiltonians?". Physical Review Letters. 72 (17): 2678–2681. doi:10.1103/PhysRevLett.72.2678.
  6. ^ Shental, O.; Kanter, I. (2009). "Shannon meets Carnot: Generalized second thermodynamic law". EPL (Europhysics Letters). 85 (1): 10006. arXiv:0806.3763. doi:10.1209/0295-5075/85/10006.
  7. ^ Kanter, Ido; Kopelowitz, Evi; Kinzel, Wolfgang (2008). "Public Channel Cryptography: Chaos Synchronization and Hilbert's Tenth Problem". Physical Review Letters. 101 (8). arXiv:0806.0931. doi:10.1103/PhysRevLett.101.084102.
  8. ^ Kanter, Ido; Aviad, Yaara; Reidler, Igor; Cohen, Elad; Rosenbluh, Michael (2010). "An optical ultrafast random bit generator". Nature Photonics. 4 (1): 58–61. Bibcode:2010NaPho...4...58K. doi:10.1038/nphoton.2009.235.
  9. ^ Sardi, Shira; Vardi, Roni; Sheinin, Anton; Goldental, Amir; Kanter, Ido (2017). "New Types of Experiments Reveal that a Neuron Functions as Multiple Independent Threshold Units". Scientific Reports. 7 (1). Springer Science and Business Media LLC. doi:10.1038/s41598-017-18363-1. PMC 5740076.
  10. ^ Sardi, Shira; Vardi, Roni; Goldental, Amir; Tugendhaft, Yael; Uzan, Herut; Kanter, Ido (2018). "Dendritic Learning as a Paradigm Shift in Brain Learning". ACS Chemical Neuroscience. 9 (6): 1230–1232. doi:10.1021/acschemneuro.8b00204.
  11. ^ "Reverberating Modes in Neural Networks | the Gonda Multidisciplinary Brain Research Center".

External links[edit]

Leave a Reply