Quantum computer benchmark Updated 2025-07-16
One important area of research and development of quantum computing is the development of benchmarks that allow us to compare different quantum computers to decide which one is more powerful than the other.
Ideally, we would like to be able to have a single number that predicts which computer is more powerful than the other for a wide range of algorithms.
However, much like in CPU benchmarking, this is a very complex problem, since different algorithms might perform differently in different architectures, making it very hard to sum up the architecture's capabilities to a single number as we would like.
The only thing that is directly comparable across computers is how two machines perform for a single algorithm, but we want a single number that is representative of many algorithms.
For example, the number of qubits would be a simple naive choice of such performance predictor number. But it is very imprecise, since other factors are also very important:
  • qubit error rate
  • coherence time, which determines the maximum circuit depth
  • qubit connectivity. Can you only connect to 4 neighbouring qubits in a 2D plane? Or to every other qubit equally as well?
Quantum volume is another less naive attempt at such metric.
One possibly interesting and possibly obvious point of view, is that a quantum computer is an experimental device that executes a quantum probabilistic experiment for which the probabilities cannot be calculated theoretically efficiently by a nuclear weapon.
This is how quantum computing was originally theorized by the likes of Richard Feynman: they noticed that "Hey, here's a well formulated quantum mechanics problem, which I know the algorithm to solve (calculate the probability of outcomes), but it would take exponential time on the problem size".
The converse is then of course that if you were able to encode useful problems in such an experiment, then you have a computer that allows for exponential speedups.
This can be seen very directly by studying one specific quantum computer implementation. E.g. if you take the simplest to understand one, photonic quantum computer, you can make systems for which you need exponential time to calculate the probabilities that photons will exit through certain holes and not others.
The obvious aspect of this idea is by coming from quantum logic gates are needed because you can't compute the matrix explicitly as it grows exponentially: knowing the full explicit matrix is impossible in practice, and knowing the matrix is equivalent to knowing the probabilities of every outcome.
Quantum computing Updated 2025-07-16
Quantum is getting hot in 2019, and even Ciro Santilli got a bit excited: quantum computing could be the next big thing.
No useful algorithm has been economically accelerated by quantum yet as of 2019, only useless ones, but the bets are on, big time.
To get a feeling of this, just have a look at the insane number of startups that are already developing quantum algorithms for hardware that doesn't/barely exists! quantumcomputingreport.com/players/privatestartup (archive). Some feared we might be in a bubble: Are we in a quantum computing bubble?
To get a basic idea of what programming a quantum computer looks like start by reading: Section "Quantum computing is just matrix multiplication".
Some people have their doubts, and that is not unreasonable, it might truly not work out. We could be on the verge of an AI winter of quantum computing. But Ciro Santilli feels that it is genuinely impossible to tell as of 2020 if something will work out or not. We really just have to try it out and see. There must have been skeptics before every single next big thing.
Quantum electrodynamics Lagrangian Updated 2025-07-16
where:
Note that this is the sum of the:
Note that the relationship between and is not explicit. However, if we knew what type of particle we were talking about, e.g. electron, then the knowledge of psi would also give the charge distribution and therefore
Video 1.
Particle Physics is Founded on This Principle! by Physics with Elliot (2022)
Source.
Quantum error correction Updated 2025-07-16
Technique that uses multiple non-ideal qubits (physical qubits) to simulate/produce one perfect qubit (logical).
One is philosophically reminded of classical error correction codes, where we also have multiple input bits per actual information bit.
TODO understand in detail. This appears to be a fundamental technique since all physical systems we can manufacture are imperfect.
Part of the fundamental interest of this technique is due to the quantum threshold theorem.
For example, when PsiQuantum raised 215M in 2020, they announced that they intended to reach 1 million physical qubits, which would achieve between 100 and 300 logical qubits.
Course outline given:
Non-relativistic QFT is a limit of relativistic QFT, and can be used to describe for example condensed matter physics systems at very low temperature. But it is still very hard to make accurate measurements even in those experiments.
Defines "relativistic" as: "the Lagrangian is symmetric under the Poincaré group".
Mentions that "QFT is hard" because (a finite list follows???):
There are no nontrivial finite-dimensional unitary representations of the Poincaré group.
But I guess that if you fully understand what that means precisely, QTF won't be too hard for you!
Notably, this is stark contrast with rotation symmetry groups (SO(3)) which appears in space rotations present in non-relativistic quantum mechanics.
Number of pages circa 2021: 155.
It should also be noted that those notes are still being updated circa 2020 much after original publication. But without Git to track the LaTeX, it is hard to be sure how much. We'll get there one day, one day.
Some quotes self describing the work:
A follow up course in the University of Cambridge seems to be the "Advanced QFT course" (AQFT, Quantum field theory II) by David Skinner: www.damtp.cam.ac.uk/user/dbs26/AQFT.html
Quantum Hall effect Updated 2025-07-16
Quantum version of the Hall effect.
As you increase the magnetic field, you can see the Hall resistance increase, but it does so in discrete steps.
Figure 1.
Hall resistance as a function of the applied magnetic field showing the Quantum Hall effect
. Source. As we can see, the blue line of the Hall resistance TODO material, temperature, etc. It is unclear if this is just
Gotta understand this because the name sounds cool. Maybe also because it is used to define the fucking ampere in the 2019 redefinition of the SI base units.
At least the experiment description itself is easy to understand. The hard part is the physical theory behind.
Video 1.
Integer and fractional quantum Hall effects by Matthew A. Grayson
. Source. Presented 2015. This dude did good.
SMIC Updated 2025-07-16
Video 1.
SMIC, Explained by Asianometry (2021)
Source.
One key insight, is that the matrix of a non-trivial quantum circuit is going to be huge, and won't fit into any amount classical memory that can be present in this universe.
This is because the matrix is exponential in the number qubits, and is more than the number of atoms in the universe!
Therefore, off the bat we know that we cannot possibly describe those matrices in an explicit form, but rather must use some kind of shorthand.
But it gets worse.
Even if we had enough memory, the act of explicitly computing the matrix is not generally possible.
This is because knowing the matrix, basically means knowing the probability result for all possible outputs for each of the possible inputs.
But if we had those probabilities, our algorithmic problem would already be solved in the first place! We would "just" go over each of those output probabilities (OK, there are of those, which is also an insurmountable problem in itself), and the largest probability would be the answer.
So if we could calculate those probabilities on a classical machine, we would also be able to simulate the quantum computer on the classical machine, and quantum computing would not be able to give exponential speedups, which we know it does.
To see this, consider that for a given input, say 000 on a 3 qubit machine, the corresponding 8-sized quantum state looks like:
000 -> 1000 0000 == (1.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0)
and therefore when you multiply it by the unitary matrix of the quantum circuit, what you get is the first column of the unitary matrix of the quantum circuit. And 001, gives the second column and so on.
As a result, to prove that a quantum algorithm is correct, we need to be a bit smarter than "just calculate the full matrix".
Which is why you should now go and read: Section "Quantum algorithm".
This type of thinking links back to how physical experiments relate to quantum computing: a quantum computer realizes a physical experiment to which we cannot calculate the probabilities of outcomes without exponential time.
So for example in the case of a photonic quantum computer, you are not able to calculate from theory the probability that photons will show up on certain wires or not.
One direct practical reason is that we need to map the matrix to real quantum hardware somehow, and all quantum hardware designs so far and likely in the future are gate-based: you manipulate a small number of qubits at a time (2) and add more and more of such operations.
While there are "quantum compilers" to increase the portability of quantum programs, it is to be expected that programs manually crafted for a specific hardware will be more efficient just like in classic computers.
TODO: is there any clear reason why computers can't beat humans in approximating any unitary matrix with a gate set?
This is analogous to what classic circuit programmers will do, by using smaller logic gates to create complex circuits, rather than directly creating one huge truth table.
The most commonly considered quantum gates take 1, 2, or 3 qubits as input.
The gates themselves are just unitary matrices that operate on the input qubits and produce the same number of output qubits.
For example, the matrix for the CNOT gate, which takes 2 qubits as input is:
1 0 0 0
0 1 0 0
0 0 0 1
0 0 1 0
The final question is then: if I have a 2 qubit gate but an input with more qubits, say 3 qubits, then what does the 2 qubit gate (4x4 matrix) do for the final big 3 qubit matrix (8x8)? In order words, how do we scale quantum gates up to match the total number of qubits?
The intuitive answer is simple: we "just" extend the small matrix with a larger identity matrix so that the sum of the probabilities third bit is unaffected.
More precisely, we likely have to extend the matrix in a way such that the partial measurement of the original small gate qubits leaves all other qubits unaffected.
For example, if the circuit were made up of a CNOT gate operating on the first and second qubits as in:
0 ----+----- 0
      |
1 ---CNOT--- 1

2 ---------- 2
then we would just extend the 2x2 CNOT gate to:
TODO lazy to properly learn right now. Apparently you have to use the Kronecker product by the identity matrix. Also, zX-calculus appears to provide a powerful alternative method in some/all cases.
Notes/book: www-thphys.physics.ox.ac.uk/people/SteveSimon/QCM2022/QuantumMatter.pdf Marked as being for Oxford MMathPhys, so it appears that this is a 4th year course normally. TODO but where is it listed under the course list of MMapthPhys? mmathphys.physics.ox.ac.uk/course-schedule
Basically the same content as: Richard Feynman Quantum Electrodynamics Lecture at University of Auckland (1979), but maybe there is some merit to this talk, as it is a bit more direct in some points. This is consistent with what is mentioned at www.feynman.com/science/qed-lectures-in-new-zealand/ that the Auckland lecture was the first attempt.
By Mill Valley, CA based producer "Sound Photosynthesis", some info on their website: sound.photosynthesis.com/Richard_Feynman.html
They are mostly a New Age production company it seems, which highlights Feynman's absolute cult status. E.g. on the last video, he's not wearing shoes, like a proper guru.
Feynman liked to meet all kinds of weird people, and at some point he got interested in the New Age Esalen Institute. Surely You're Joking, Mr. Feynman this kind of experience a bit, there was nude bathing on a pool that oversaw the sea, and a guy offered to give a massage to the he nude girl and the accepted.
youtu.be/rZvgGekvHest=5105 actually talks about spin, notably that the endpoint events also have a spin, and that the transition rules take spin into account by rotating thing, and that the transition rules take spin into account by rotating things.
Real-time attack speedrun Updated 2025-07-16
Ciro Santilli views humans as biological robots, and therefore RTA videos can be thought of as probabilistic TAS with human achievable reflex constraints.
This aspect is especially highlighted in "speed run record evolution videos", which can be quite fun, e.g. www.youtube.com/watch?v=pmS9e7kzgS4 Ocarina of Time - World Record History and Progression (Any% Speedrun, 1990s-2017) by retro (2017)
From a similar point of view, Ciro also sometimes watches/learns a bit about competitive PvP games from a "could a computer play this better than a human" point of view.
Ciro also likes to watch commented manual speedruns of games as a way of experiencing the game at a high level without spending too much time on it, often from Games Done Quick. Their format is good because it generally showcases one player focusing more on the gameplay, and three couch commentators to give context, that's a good setup.
It is a

There are unlisted articles, also show them or only show them.