Analog computer Updated +Created
Some of the earlier computers of the 20th centure were analog computers, not digital.
At some point analog died however, and "computer" basically by default started meaning just "digital computer".
As of the 2010's and forward, with the limit of Moore's law and the rise of machine learning, people have started looking again into analog computing as a possile way forward. A key insight is that huge floating point precision is not that crucial in many deep learning applications, e.g. many new digital designs have tried 16-bit floating point as opposed to the more traditional 32-bit minium. Some papers are even looking into 8-bit: dl.acm.org/doi/10.5555/3327757.3327866
As an example, the Lightmatter company was trying to implement silicon photonics-based matrix multiplication.
A general intuition behind this type of development is that the human brain, the holy grail of machine learning, is itself an analog computer.
Silicon photonics Updated +Created
Video 1.
Silicon Photonics: The Next Silicon Revolution? by Asianometry (2022)
Source.
Video 2.
Running Neural Networks on Meshes of Light by Asianometry (2022)
Source.
Video 3.
Silicon Photonics for Extreme Computing by Keren Bergman (2017)
Source.