Analog computer Updated +Created
Some of the earlier computers of the 20th centure were analog computers, not digital.
At some point analog died however, and "computer" basically by default started meaning just "digital computer".
As of the 2010's and forward, with the limit of Moore's law and the rise of machine learning, people have started looking again into analog computing as a possile way forward. A key insight is that huge floating point precision is not that crucial in many deep learning applications, e.g. many new digital designs have tried 16-bit floating point as opposed to the more traditional 32-bit minium. Some papers are even looking into 8-bit: dl.acm.org/doi/10.5555/3327757.3327866
As an example, the Lightmatter company was trying to implement silicon photonics-based matrix multiplication.
A general intuition behind this type of development is that the human brain, the holy grail of machine learning, is itself an analog computer.
Continuous-variable quantum information Updated +Created
TODO synonym to analog quantum computer?
It is also possible to carry out quantum computing without qubits using processes with a continuous spectrum of measurement.
As of 2020, these approaches seem less developed/promising, but who knows.
These computers can be seen as analogous to classical non-quantum analog computers.
Silicon photonics Updated +Created
Video 1.
Silicon Photonics: The Next Silicon Revolution? by Asianometry (2022)
Source.
Video 2.
Running Neural Networks on Meshes of Light by Asianometry (2022)
Source.
Video 3.
Silicon Photonics for Extreme Computing by Keren Bergman (2017)
Source.