• 0 Posts
  • 41 Comments
Joined 7 months ago
cake
Cake day: July 7th, 2024

help-circle
  • Interesting you get downvoted for this when I mocked someone for saying the opposite who claimed that $0.5m was some enormous amount of money we shouldn’t be wasting, and I simply pointed out that we waste literally billions around the world on endless wars killing random people for now reason, so it is silly to come after small bean quantum computing if budgeting is your actual concern. People seemed to really hate me for saying that, or maybe it was because they just actually like wasting moneys on bombs to drop on children and so they want to cut everything but that.






  • We don’t know what it is. We don’t know how it works. That is why

    If you cannot tell me what you are even talking about then you cannot say “we don’t know how it works,” because you have not defined what “it” even is. It would be like saying we don’t know how florgleblorp works. All humans possess florgleblorp and we won’t be able to create AGI until we figure out florgleblorp, then I ask wtf is florgleblorp and you tell me “I can’t tell you because we’re still trying to figure out what it is.”

    You’re completely correct. But you’ve gone on a very long rant to largely agree with the person you’re arguing against.

    If you agree with me why do you disagree with me?

    Consciousness is poorly defined and a “buzzword” largely because we don’t have a fucking clue where it comes from, how it operates, and how it grows.

    You cannot say we do not know where it comes from if “it” does not refer to anything because you have not defined it! There is no “it” here, “it” is a placeholder for something you have not actually defined and has no meaning. You cannot say we don’t know how “it” operates or how “it” grows when “it” doesn’t refer to anything.

    When or if we ever define that properly

    No, that is your first step, you have to define it properly to make any claims about it, or else all your claims are meaningless. You are arguing about the nature of florgleblorp but then cannot tell me what florgleblorp is, so it is meaningless.

    This is why “consciousness” is interchangeable with vague words like “soul.” They cannot be concretely defined in a way where we can actually look at what they are, so they’re largely irrelevant. When we talk about more concrete things like intelligence, problem-solving capabilities, self-reflection, etc, we can at least come to some loose agreement of what that looks like and can begin to have a conversation of what tests might actually look like and how we might quantify it, and it is these concrete things which have thus been the basis of study and research and we’ve been gradually increasing our understanding of intelligent systems as shown with the explosion of AI, albeit it still has miles to go.

    However, when we talk about “consciousness,” it is just meaningless and plays no role in any of the progress actually being made, because nobody can actually give even the loosest iota of a hint of what it might possibly look like. It’s not defined, so it’s not meaningful. You have to at least specify what you are even talking about for us to even begin to study it. We don’t have to know the entire inner workings of a frog to be able to begin a study on frogs, but we damn well need to be able to identify something as a frog prior to studying it, or else we would have no idea that the thing we are studying is actually a frog.

    You cannot study anything without being able to identify it, which requires defining it at least concretely enough that we can agree if it is there or not, and that the thing we are studying is actually the thing we aim to study. We should I believe your florgleblorp, sorry, I mean “consciousness” you speak of, even exists if you cannot even tell me how to identify it? It would be like if someone insisted there is a florgleblorp hiding in my room. Well, I cannot distinguish between a room with or without a florgleblorp, so by Occam’s razor I opt to disbelieve in its existence. Similarly, if you cannot tell me how to distinguish between something that possesses this “consciousness” and something that does not, how to actually identify it in reality, then by Occam’s razor I opt to disbelieve in its existence.

    It is entirely backwards and spiritualist thinking that is popularized by all the mystics to insist that we need to study something they cannot even specify what it is first in order to figure out what it is later. That is the complete reversal of how anything works and is routinely used by charlatans to justify pseudoscientific “research.” You have to specify what it is being talked about first.


  • we need to figure out what consciousness is

    Nah, “consciousness” is just a buzzword with no concrete meaning. The path to AGI has no relevance to it at all. Even if we develop a machine just as intelligent as human beings, maybe even moreso, that can solve any arbitrary problem just as efficiently, mystics will still be arguing over whether or not it has “consciousness.”

    Edit: You can downvote if you want, but I notice none of you have any actual response to it, because you ultimately know it is correct. Keep downvoting, but not a single one of you will actually reply and tell us me how we could concretely distinguish between something that is “conscious” and something that isn’t.

    Even if we construct a robot that fully can replicate all behaviors of a human, you will still be there debating over whether or not is “conscious” because you have not actually given it a concrete meaning so that we can identify if something actually has it or not. It’s just a placeholder for vague mysticism, like “spirit” or “soul.”

    I recall a talk from Daniel Dennett where he discussed an old popular movement called the “vitalists.” The vitalists used “life” in a very vague meaningless way as well, they would insist that even if understand how living things work mechanically and could reproduce it, it would still not be considered “alive” because we don’t understand the “vital spark” that actually makes it “alive.” It would just be an imitation of a living thing without the vital spark.

    The vitalists refused to ever concretely define what the vital spark even was, it was just a placeholder for something vague and mysterious. As we understood more about how life works, vitalists where taken less and less serious, until eventually becoming largely fringe. People who talk about “consciousness” are also going to become fringe as we continue to understand neuroscience and intelligence, if scientific progress continues, that is. Although this will be a very long-term process, maybe taking centuries.



  • The space mechanics was definitely one of the great things about that game, in my opinion. Most space games when you land you just press a button and it plays an animation. Having to land manually with a landing camera is very satisfying. When you crash and parts of your ship break and you have to float outside to fix it, that was also very fun. I feel like a lot of space games are a bit lazy about the actual space mechanics, this game did it very well.


  • Complex numbers are just a way of representing an additional degree of freedom in an equation. You have to represent complex numbers not on a number line but on the complex plane, so each complex number is associated with two numbers. That means if you create a function that requires two inputs and two outputs, you could “compress” that function into a single input and output by using complex numbers.

    Complex numbers are used all throughout classical mechanics. Waves are two-dimensional objects because they both have an amplitude and a wavelength. Classical wave dynamics thus very often use complex numbers because you can capture the properties of waves more concisely. An example of this is the Fourier transform. If you look up the function, it looks very scary, it has an integral and Euler’s number raised to the negative power of the imaginary number multiplied by pi. However, if you’ve worked with complex numbers a lot, you’d immediately recognize that raising Euler’s number to pi times the imaginary number is just how you represent rotations on the complex plane.

    Despite how scary the Fourier transform looks, literally all it is actually doing is wrapping a wave around a circle. 3Blue1Brown has a good video on his channel of how to visualize the Fourier transform. The Fourier transform, again, isn’t inherently anything quantum mechanical, we use it all the time in classical mechanics, for example, if you ever used an old dial-up model and wondered why it made those weird noises, it was encoding data as sound wave by representing them as different harmonic waves that it would then add together, producing that sound. The Fourier transform could then be used by the modem at the other end to break the sound back apart into those harmonic waves and then decode it back into data.

    In quantum mechanics, properties of systems always have an additional kind of “orientation” to them. When particles interact, if their orientations are aligned, the outcome of the interaction is deterministic. If they are misaligned, then it introduces randomness. For example, an electron’s spin state can either be up or down. However, its spin state also has a particular orientation to it, so you can only measure it “correctly” by having the orientation of the measuring device aligned with the electron. If they are misaligned, you introduce randomness. These orientations often are associated with physical rotations, for example, with the electron spins state, you measure it with something known as a Stern-Gerlach apparatus, and to measure the electron on a different orientation you have to physically rotate the whole apparatus.

    Because the probability of measuring certain things directly relates to the relative orientation between your measuring device and the particle, it would be nice if we had a way to represent both the relative orientation and the probability at the same time. And, of course, you guessed it, we do. It turns out you can achieve this simply by representing your probability amplitudes (the % chance of something occurring) as complex numbers. This means in quantum mechanics, for example, an event can have a -70.7i% chance of occurring.

    While that sounds weird at first, you quickly realize that the only reason we represent it this way is because it directly connects the relative orientation between the systems interacting and the probabilities of certain outcomes. You see, you can convert quantum probabilities to classical just by computing the distance from 0% on the complex plane and squaring it, which in the case of -70.7i% would give you 50%, which tells you this just means it is basically a fair coin flip. However, you can also compute from this number the relative orientation of the two measuring devices, which in this case you would find it to be rotated 90 degrees. Hence, because both values can be computed from the same number, if you rotate the measuring device it must necessarily alter the probabilities of different outcomes.

    You technically don’t need to ever use complex numbers. You could, for example, take the Schrodinger equation and just break it up into two separate equations for the real and imaginary part, and have them both act on real numbers. Indeed, if you actually build a quantum computer simulator in a classical computer, most programming languages don’t include complex numbers, so all your algorithms have to break the complex numbers into two real numbers. It’s just when you are writing down these equations, they can get very messy this way. Complex numbers are just far more concise to represent additional degrees of freedom without needing additional equations/functions.


  • This is a rather reductive view of quantum cryptography.

    Correct = reductive?

    The two most common applications of it I hear about is the development of encryption algorithms resistant to being broken on quantum computers

    First, I was talking about quantum encryption, not quantum cryptography, which is a bit more broad. Second, we already have cryptographic algorithms that run on classical computers that are not crackable by quantum computers, known as lattice-based cryptography which are way more practical than anything quantum cryptography could offer.

    the way, say, Shur’s algorithm is known to break RSA

    Shor’s algorithm. Yes, it breaks asymmetrical ciphers like RSA, but we have developed alternatives already it cannot break, like Kyber.

    and techniques like quantum key distribution

    Classical key exchange algorithms prevent someone from reading your key if they intercept the data packets between you. QKD is entirely impractical because it does not achieve this. Rather than preventing someone from reading your key if they intercept the data packets, it merely allows you to detect if someone is intercepting the data packets. You see, in regular cryptography, you want people to be able to intercept your data. It’s necessary for something like the internet to work, because packets of data have to be passed around the whole world, and it would suck if your packets got lost simply because someone read them in transit, which is why QKD is awful. If a single person reads the data packet in transit then they would effectively deny service to the recipient.

    Both of these are real problems that don’t become meaningless just because one-time pads exist - you need to somehow securely distribute the keys for one-time-pad encryption.

    One-time pad encryption is awful as I already explained, it would cut the entire internet bandwidth in half because if you wanted to transmit 10 gigabytes of data you would also need to transmit 10 gigabyte key. QKD is also awful for the fact that it would be unscalable to an “internet” because of how easy it is to deny service. It also doesn’t even guarantee you can detect someone snooping your packets because it is susceptible to a man-in-the-middle attack. Sure, the Diffie-Hellman Key Exchange is also susceptible to a man-in-the-middle attack, but we solve this using public key infrastructure. You cannot have public key infrastructure for quantum cryptography.

    The only proposed quantum digital signature algorithms are unscalable because they rely on Holevo’s theorem, which basically says there is a limited amount of information about the quantum state of a qubit you can gather from a single measurement, thus creating a sort of one-way function that can be used for digital signatures. The issue with this is that Holevo’s theorem also says you can acquire more information if you have more copies of the same qubit, i.e. it means every time you distribute a copy of the public key, you increase the probability someone could guess it. Public keys would have to be consumable which would entirely prevent you from scaling it to any significantly large network.

    That’s why one-time pads aren’t used everywhere, (“it would cut the whole internet bandwidth in half overnight” would not have been a sufficient reason - that’d be a tiny price to pay for unbreakable encryption, if it actually worked).

    You are living in fairy tale lala land. Come back down to reality. If you offer someone an algorithm that is impossible to break in a trillion, trillion years, and another algorithm that is in principle impossible to break, but the former algorithm is twice as efficient, then every company on the entirety of planet earth will choose the former. No enterprise on earth is going to double their expenses for something entirely imaginary that could never be observed in practice. You are really stuck in delulu town if you unironically think the reason one-time pads aren’t used practically is due to lack of secure key distribution.

    Even prior to the discovery of Shor’s algorithm, we were issuing DHKE which, at the time, was believed to be pretty much an unbreakable way to share keys. Yet, even in this time before people knew DHKE could be potentially broken by quantum computers, nobody used DHKE to exchange keys for one-time pads. DHKE is always used to exchange keys for symmetrical ciphers like AES. AES256 is not breakable by quantum computers in practice as even a quantum computer would require trillions of years to break it. There is zero reason to use a one-time pad when something like AES exists. It’s the industry standard for a reason and I bet you my entire life savings we are not going to abandon it for one-time pads ever.



  • Quantum encryption won’t ever be a “thing.”

    All cryptography requires a pool of random numbers as inputs, and while different cryptographic methods are more secure than others, all of them are only as secure as their random number pool. The most secure cipher possible is known as a one-time pad which can be proven to be as secure as a cryptographic algorithm could possibly be, and so the only thing that could possibly lead to it being hacked is a poor random number pool. Since quantum mechanics can be used to generate truly random numbers, you could have a perfect random number pool, combined with a perfect cipher, gives you perfect encryption.

    That sounds awesome right? Well… no. Because it is trivially easy these days to get regular old classical computers to spit out basically an indefinite number of pseudorandom numbers that are indistinguishable from truly random numbers. Why do you think modern operating systems allow you to encrypt your whole drive? You can have a file tens of gigabytes bit and you click it and it opens instantly, despite your whole drive being encrypted, because your CPU can generate tens of gigabytes of random numbers good enough for cryptography faster than you can even blink.

    Random number generation is already largely a solved problem for classical computers. I own a quantum random number generator. I can compare it in various test suites such as the one released by NIST to test the quality of a random number generator, and it can’t tell the different between that and my CPU’s internal random number generator. Yes, the CPU. Most modern CPUs both have the ability to collect entropy data from thermal noise to seed a pseudorandom number generator, as well as having a hardware-level pseudorandom number, such as x86’s RDSEED and RDRAND instructions, so they can generate random numbers good enough for cryptography at blazing speeds.

    The point is that in practice you will never actually notice, even if you were a whole team of PhD statisticians and mathematicians, the difference between a message encrypted by a quantum computer and a message encrypted by a classical computer using an industry-approved library. Yet, it is not just that they’re equal, quantum encryption would be far worse. We don’t use one-time pads in practice despite their security because they require keys as long as the message itself, and thus if we adopted them, it would cut the whole internet bandwidth in half overnight. Pseudorandom number generators are superior to use as the basis for cryptography because the key can be very small and then it can spit out the rest of what is needed to encrypt/decrypt the message from it, and deterministic encryption/decryption algorithms like AES and ChaCha20 are not crackable even by a quantum computer.


  • Honestly, the random number generation on quantum computers is practically useless. Speeds will not get anywhere near as close to a pseudorandom number generator, and there are very simple ones you can implement that are blazing fast, far faster than any quantum computer will spit out, and produce numbers that are widely considered in the industry to be cryptographically secure. You can use AES for example as a PRNG and most modern CPUs like x86 processor have hardware-level AES implementation. This is why modern computers allow you to encrypt your drive, because you can have like a file that is a terabyte big that is encrypted but your CPU can decrypt it as fast as it takes for the window to pop up after you double-click it.

    While PRNG does require an entropy pool, the entropy pool does not need to be large, you can spit out terabytes of cryptographically secure pseudorandom numbers on a fraction of a kilobyte of entropy data, and again, most modern CPUs actually include instructions to grab this entropy data, such as Intel’s CPUs have an RDSEED instruction which let you grab thermal noise from the CPU. In order to avoid someone discovering a potential exploit, most modern OSes will mix into this pool other sources as well, like fluctuations in fan voltage.

    Indeed, used to with Linux, you had a separate way to read random numbers directly from the entropy pool and another way to read pseudorandom numbers, those being /dev/random and /dev/urandom. If you read from the entropy pool, if it ran out, the program would freeze until it could collect more, so some old Linux programs you would see the program freeze until you did things like move your mouse around.

    But you don’t see this anymore because generating enormous amounts of cryptographysically secure random nubmers is so easy with modern algorithms that modern Linux just collects a little bit of entropy at boot and it uses that to generate all pseudorandom numbers after, and just got rid of needing to read it directly, both /dev/random and /dev/urandom now just internally in the OS have the same behavior. Any time your PC needs a random number it just pulls from the pseudorandom number generator that was configured at boot, and you have just from the short window of collecting entropy data at boot the ability to generate sufficient pseudorandom numbers basically forever, and these are the numbers used for any cryptographic application you may choose to run.

    The point of all this is to just say random number generation is genuinely a solved problem, people don’t get just how easy it is to basically produce practically infinite cryptographically secure pseudorandom numbers. While on paper quantum computers are “more secure” because their random numbers would be truly random, in practice you literally would never notice a difference. If you gave two PhD mathematicians or statisticians the same message, one encrypted using a quantum random number generator and one encrypted with a PRNG like AES or ChaCha20, and asked them to decipher them, they would not be able to decipher either. In fact, I doubt they would even be able to identify which one was even encoded using the quantum random number generator. A string of random numbers looks just as “random” to any random number test suite whether or not it came from a QRNG or a high-quality PRNG (usually called CSPRNG).

    I do think at least on paper quantum computers could be a big deal if the engineering challenge can ever be overcome, but quantum cryptography such as “the quantum internet” are largely a scam. All the cryptographic aspects of quantum computers are practically the same, if not worse, than traditional cryptography, with only theoretical benefits that are technically there on paper but nobody would ever notice in practice.


  • the study that found the universe is not locally real. Things only happen once they are observed

    This is only true if you operate under a very specific and strict criterion of “realism” known as metaphysical realism. Einstein put forward a criterion of what he thought this philosophy implied for a physical theory, and his criterion is sometimes called scientific realism.

    Metaphysical realism is a very complex philosophy. One of its premises is that there exists an “absolute” reality where all objects are made up of properties that are independent of perspective. Everything we perceive is wholly dependent upon perspective, so metaphysical realism claims that what we perceive is not “true” reality but sort of an illusion created by the brain. “True” reality is then treated as the absolute spacetime filled with particles captured in the mathematics of Newton’s theory.

    The reason it relies on this premise is because by assigning objects perspective invariant properties, then they can continue to exist even if no other object is interacting with them, or, more specifically, they continue to exist even if “no one is looking at them.” For example, if you fire a cannonball from point A to point B, and you only observe it leaving point A and arriving at point B, Newtonian mechanics allows you to “track” its path between these two points even if you did not observe it.

    The problem is that you cannot do this in quantum mechanics. If you fire a photon from point A to point B, the theory simply disallows you from unambiguously filling in the “gaps” between the two points. People then declare that “realism is dead,” but this is a bit misleading because this is really only a problem for metaphysical/scientific realism. There are many other kinds of realism in literature.

    For example, the philosopher Jocelyn Benoist’s contextual realism argues that the exact opposite. The mathematical theory is not “true reality” but is instead a description of reality. A description of reality is not the same as reality. Would a description of the Eiffel Tower substitute actually seeing it in reality? Of course not, they’re not the same. Contextual realism instead argues that what is real is not the mathematical description but is precisely what we perceive. The reason we perceive reality in a way that depends upon perspective is because reality is just relative (or “contextual”). There is no “absolute” reality but only a contextual reality and that contextual reality we perceive directly as it really is.

    Thus for contextual realism, there is no issue with the fact that we cannot “track” things unambiguously, because it has no attachment to treating particles as if they persist as autonomous entities. It is perfectly fine with just treating it as if the particle hops from point A to point B according to some predictable laws and relative to the context in which the observer occupies. That is just how objective reality works. Observation isn’t important, and indeed, not even measurement, because whatever you observe in the experimental setting is just what reality is like in that context. The only thing that “arises” is your identification.


  • Why did physicists start using the word “real” and “realism”? It’s a philosophical term, not a physical one, and it leads to a lot of confusion. “Local” has a clear physical meaning, “realism” gets confusing. I have seen some papers that use “realism” in a way that has a clear physical definition, such as one I came across defined it in terms of a hidden variable theory. Yet, I also saw a paper coauthored by the great Anton Zeilinger that speaks of “local realism,” but very explicitly uses “realism” with its philosophical meaning, that there is an objective reality independent of the observer, which to me it is absurd to pretend that physics in any way calls this into account.

    If you read John Bell’s original paper “On the Einstein Podolsky Rosen Paradox,” he never once use the term “realism.” The only time I have seen “real” used at all in this early discourse is in the original EPR paper, but this was merely a “criterion” (meaning a minimum but not sufficient condition) for what would constitute a theory that is a complete description of reality. Einstein/Podolsky/Rosen in no way presented this as a definition of “reality” or a kind of “realism.”

    Indeed, even using the term “realism” on its own is ambiguous, as there are many kinds of “realisms” in the literature. The phrase “local realism” on its own is bound to lead to confusion, and it does, because I pointed out, even in the published literature physicists do not always use “realism” consistently. If you are going to talk about “realism,” you need to preface it to be clear what kind of realism you are specifically talking about.

    If the reason physicists started to talk about “realism” is because they specifically are referring to something that includes the EPR criterion, then they should call it “EPR realism” or something like that. Just saying “realism” is so absurdly ridiculous it is almost as if they are intentionally trying to cause confusion. I don’t really blame anyone who gets confused on this because like I said if you even read the literature there is not even consistent usage in the peer-reviewed papers.

    The phrase “observer-dependence” is also very popular in the published literature. So, while I am not disagreeing with you that “observation” is just an interaction, this is actually a rather uncommon position known as relational quantum mechanics.


  • A lot of people who present quantum mechanics to a laymen audience seem to intentionally present it to be as confusing as possible because they like the “mystery” behind it. Yet, it is also easy to present it in a trivially simple and boring way that is easy to understand.

    Here, I will tell you a simple framework that is just 3 rules and if you keep them in mind then literally everything in quantum mechanics makes sense and follows quite simply.

    1. Quantum mechanics is a probabilistic theory where, unlike classical probability theory, the probabilities of events can be complex-valued. For example, it is meaningful in quantum mechanics for an event to have something like a -70.7i% chance of occurring.
    2. The physical interpretation of complex-valued probabilities is that the further the probability is from zero, the more likely it is. For example, an event with a -70.7i% probability of occurring is more likely than one with a 50% probability of occurring because it is further from zero. (You can convert quantum probabilities to classical just by computing their square magnitudes, which is known as the Born rule.)
    3. If two events or more become statistically correlated with one another (this is known as “entanglement”) the rules of quantum mechanics disallows you from assigning quantum probabilities to the individual systems taken separately. You can only assign the quantum probabilities to the two events or more taken together. (The only way to recover the individual probabilities is to do something called a partial trace to compute the reduced density matrix.)

    If you keep those three principles in mind, then everything in quantum mechanics follows directly, every “paradox” is resolved, there is no confusion about anything.

    For example, why is it that people say quantum mechanics is fundamentally random? Well, because if the universe is deterministic, then all outcomes have either a 0% or 100% probability, and all other probabilities are simply due to ignorance (what is called “epistemic”). Notice how 0% and 100% have no negative or imaginary terms. They thus could not give rise to quantum effects.

    These quantum effects are interference effects. You see, if probabilities are only between 0% and 100% then they can only be cumulative. However, if they can be negative, then the probabilities of events can cancel each other out and you get no outcome at all. This is called destructive interference and is unique to quantum mechanics. Interference effects like this could not be observed in a deterministic universe because, in reality, no event could have a negative chance of occurring (because, again, in a deterministic universe, the only possible probabilities are 0% or 100%).

    If we look at the double-slit experiment, people then ask why does the interference pattern seem to go away when you measure which path the photon took. Well, if you keep this in mind, it’s simple. There’s two reasons actually and it depends upon perspective.

    If you are the person conducting the experiment, when you measure the photon, it’s impossible to measure half a photon. It’s either there or it’s not, so 0% or 100%. You thus force it into a definite state, which again, these are deterministic probabilities (no negative or imaginary terms), and thus it loses its ability to interfere with itself.

    Now, let’s say you have an outside observer who doesn’t see your measurement results. For him, it’s still probabilistic since he has no idea which path it took. Yet, the whole point of a measuring device is to become statistically correlated with what you are measuring. So if we go to rule #3, the measuring device should be entangled with the particle, and so we cannot apply the quantum probabilities to the particle itself, but only to both the particle and measuring device taken together.

    Hence, for the outside observer’s perspective, only the particle and measuring device collectively could exhibit quantum interference. Yet, only the particle passes through the two slits on its own, without the measuring device. Thus, they too would predict it would not interfere with itself.

    Just keep these three rules in mind and you basically “get” quantum mechanics. All the other fluff you hear is people attempting to make it sound more mystical than it actually is, such as by interpreting the probability distribution as a literal physical entity, or even going more bonkers and calling it a grand multiverse, and then debating over the nature of this entity they entirely made up.

    It’s literally just statistics with some slightly different rules.




  • Why are you isolating a single algorithm? There are tons of them that speed up various aspects of linear algebra and not just that single one, and many improvements to these algorithms since they were first introduced, there are a lot more in the literature than just in the popular consciousness.

    The point is not that it will speed up every major calculation, but these are calculations that could be made use of, and there will likely even be more similar algorithms discovered if quantum computers are more commonplace. There is a whole branch of research called quantum machine learning that is centered solely around figuring out how to make use of these algorithms to provide performance benefits for machine learning algorithms.

    If they would offer speed benefits, then why wouldn’t you want to have the chip that offers the speed benefits in your phone? Of course, in practical terms, we likely will not have this due to the difficulty and expense of quantum chips, and the fact they currently have to be cooled below to near zero degrees Kelvin. But your argument suggests that if somehow consumers could have access to technology in their phone that would offer performance benefits to their software that they wouldn’t want it.

    That just makes no sense to me. The issue is not that quantum computers could not offer performance benefits in theory. The issue is more about whether or not the theory can be implemented in practical engineering terms, as well as a cost-to-performance ratio. The engineering would have to be good enough to both bring the price down and make the performance benefits high enough to make it worth it.

    It is the same with GPUs. A GPU can only speed up certain problems, and it would thus be even more inefficient to try and force every calculation through the GPU. You have libraries that only call the GPU when it is needed for certain calculations. This ends up offering major performance benefits and if the price of the GPU is low enough and the performance benefits high enough to match what the consumers want, they will buy it. We also have separate AI chips now as well which are making their way into some phones. While there’s no reason at the current moment to believe we will see quantum technology shrunk small and cheap enough to show up in consumer phones, if hypothetically that was the case, I don’t see why consumers wouldn’t want it.

    I am sure clever software developers would figure out how to make use of them if they were available like that. They likely will not be available like that any time in the near future, if ever, but assuming they are, there would probably be a lot of interesting use cases for them that have not even been thought of yet. They will likely remain something largely used by businesses but in my view it will be mostly because of practical concerns. The benefits of them won’t outweigh the cost anytime soon.