That’s because the capital-s Singularity as proposed by Verner Vinge is what we’re worried about here. The advent of a technological achievement that forever changes humanity, possibly signalling the end of it.
This does specifically set a barrier, which is a “Point of No Return” when it comes to technology. This is what most people mean when they mean the Singularity. When a program becomes capital-I Intelligent.
Neumann’s original proposal is as limited by mathematics as an LLM itself. The term Singularity has, as is common in the English language, become a larger term to signify a barrier has been crossed. There are other theories beyond the idea that it’s just self-replication gone wild.
You’re trying to reduce what to most people is a moral quandry to pure mathematics. Since my core point is that pure mathematics is not enough to capture the depth and potential of humanity, I’m not going to be swayed by being told it’s just a mathematical function.
I will give you a boost for being interesting, though.
I think the issue here is you’re interpolating a couple different concepts:
Iterated technological self-improvement resulting in exponential growth
Artificial General Intelligence
The threat to humanity from advanced AI
1 is the singularity, 2 and 3 are frequently hypothesized consequences of 1. Kinda like extensive use of fossil fuels is one concept, the greenhouse effect is another, and rising sea levels a third. They are related, but distinct, even though one contributes to another.
Combining related concepts under one term dilutes the term and makes it more difficult to effectively communicate. Of course, the moral quandaries are valuable topics of discussion, but the mathematical function is a separate topic, and likewise valuable in and of itself
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
We only have to worry about machine overlords if we PUT machines in charge of stuff, and we’ll only do that if we think they are intelligent enough to make decisions. So yeah, determining whether it has real intelligent is a key thing here.
Again, you are hung up on semantics and terminology. You are going down a checklist based on one specific person’s extrapolation on the possible consequences of the implementation of a concept. I am looking at the core concept underlying that extrapolation (the exponential increase in capability of a system, due to the recursive application of the system’s transformative capabilities to the architecture underlying those same capabilities).
You are caught up on whether the ability to operate on the basis of more data every second than any human can digest in an academic lifetime qualifies as “superhuman”. You are hung up on the same extraneous and irrelevant concepts you introduced: consciousness, accountability, decisions, understanding, inspiration.
My original statement was that the singularity doesn’t look like the singularity until it does.
Even your liberal definitions still rotate around the concept of exponential iterative growth (despite their addition of functionally extraneous {though derivative} concepts like supremacy or emergent consciousness). There’s nothing more that I can say there. You’re going on about definitions changing, the center of the definition is the same. Iteration. Self-programming. Exponential growth.
It doesn’t look like it until it does. That’s what the exponential function does. It’s nearly horizontal, negligable, barely noticable gradual growth; until it hits the anchor point when it rockets up, nearly vertical, almost infinite growth. That’s the core concept at play. Learning to crawl for months, then setting impossible records the next day.
Learn what an exponential function is. Learn why it looks like that, and what the anchor point represents. Learn how LLMs work. Look into Microsoft’s LongNet.
It’s not going to look like the singularity, right up until it does
@agamemnonymous Take it up with Verner, man. The idea’s been popularized in a way that gathers all three, and there’s even theories about a Non-AI Singularity.
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
We only have to worry about machine overlords if we PUT machines in charge of stuff, and we’ll only do that if we think they are intelligent enough to make decisions. So yeah, determining whether it has real intelligent is a key thing here.
Popularity is not correctness. You’re using a sloppily defined term. I’m using the fundamental definition. Your (Verners) concept muddles matters pointlessly.
The fact is, self-refining LLMs can very possibly exhibit the intelligence explosion fundamental to Von Neumann or I.J. Good’s definition. They are already beginning to alter the way human society operates (coding, school, replacing jobs). They easily pass the Turing test with the right prompts. Your whole point is that it’s not “real” intelligence because they don’t really “understand”, but I can say the same for you. For all I know, you’re an LLM and there’s literally no way that you can prove you aren’t.
Lines in the sand about “real” intelligence are purely philosophical, and that kind of hyperopic philosophizing is exactly the sort of behavior that dooms humanity via underestimation. I’d rather we didn’t find ourselves under machine overlords because “technically they aren’t even really intelligent”.
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
We only have to worry about machine overlords if we PUT machines in charge of stuff, and we’ll only do that if we think they are intelligent enough to make decisions. So yeah, determining whether it has real intelligent is a key thing here.
I’m invested in nothing, there is no con. I’m sorry, you do not seem to understand the fundamental concepts at play. I would recommend trying to learn but I understand if you cannot.
That’s because the capital-s Singularity as proposed by Verner Vinge is what we’re worried about here. The advent of a technological achievement that forever changes humanity, possibly signalling the end of it.
This does specifically set a barrier, which is a “Point of No Return” when it comes to technology. This is what most people mean when they mean the Singularity. When a program becomes capital-I Intelligent.
Neumann’s original proposal is as limited by mathematics as an LLM itself. The term Singularity has, as is common in the English language, become a larger term to signify a barrier has been crossed. There are other theories beyond the idea that it’s just self-replication gone wild.
You’re trying to reduce what to most people is a moral quandry to pure mathematics. Since my core point is that pure mathematics is not enough to capture the depth and potential of humanity, I’m not going to be swayed by being told it’s just a mathematical function.
I will give you a boost for being interesting, though.
I think the issue here is you’re interpolating a couple different concepts:
Iterated technological self-improvement resulting in exponential growth
Artificial General Intelligence
The threat to humanity from advanced AI
1 is the singularity, 2 and 3 are frequently hypothesized consequences of 1. Kinda like extensive use of fossil fuels is one concept, the greenhouse effect is another, and rising sea levels a third. They are related, but distinct, even though one contributes to another.
Combining related concepts under one term dilutes the term and makes it more difficult to effectively communicate. Of course, the moral quandaries are valuable topics of discussion, but the mathematical function is a separate topic, and likewise valuable in and of itself
@agamemnonymous
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
We only have to worry about machine overlords if we PUT machines in charge of stuff, and we’ll only do that if we think they are intelligent enough to make decisions. So yeah, determining whether it has real intelligent is a key thing here.
(Dammit, we’ve reached the end of the chat tree)
Again, you are hung up on semantics and terminology. You are going down a checklist based on one specific person’s extrapolation on the possible consequences of the implementation of a concept. I am looking at the core concept underlying that extrapolation (the exponential increase in capability of a system, due to the recursive application of the system’s transformative capabilities to the architecture underlying those same capabilities).
You are caught up on whether the ability to operate on the basis of more data every second than any human can digest in an academic lifetime qualifies as “superhuman”. You are hung up on the same extraneous and irrelevant concepts you introduced: consciousness, accountability, decisions, understanding, inspiration.
My original statement was that the singularity doesn’t look like the singularity until it does.
Even your liberal definitions still rotate around the concept of exponential iterative growth (despite their addition of functionally extraneous {though derivative} concepts like supremacy or emergent consciousness). There’s nothing more that I can say there. You’re going on about definitions changing, the center of the definition is the same. Iteration. Self-programming. Exponential growth.
It doesn’t look like it until it does. That’s what the exponential function does. It’s nearly horizontal, negligable, barely noticable gradual growth; until it hits the anchor point when it rockets up, nearly vertical, almost infinite growth. That’s the core concept at play. Learning to crawl for months, then setting impossible records the next day.
Learn what an exponential function is. Learn why it looks like that, and what the anchor point represents. Learn how LLMs work. Look into Microsoft’s LongNet.
It’s not going to look like the singularity, right up until it does
@agamemnonymous Take it up with Verner, man. The idea’s been popularized in a way that gathers all three, and there’s even theories about a Non-AI Singularity.
This happens all the time with terms.
@agamemnonymous
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
We only have to worry about machine overlords if we PUT machines in charge of stuff, and we’ll only do that if we think they are intelligent enough to make decisions. So yeah, determining whether it has real intelligent is a key thing here.
Popularity is not correctness. You’re using a sloppily defined term. I’m using the fundamental definition. Your (Verners) concept muddles matters pointlessly.
The fact is, self-refining LLMs can very possibly exhibit the intelligence explosion fundamental to Von Neumann or I.J. Good’s definition. They are already beginning to alter the way human society operates (coding, school, replacing jobs). They easily pass the Turing test with the right prompts. Your whole point is that it’s not “real” intelligence because they don’t really “understand”, but I can say the same for you. For all I know, you’re an LLM and there’s literally no way that you can prove you aren’t.
Lines in the sand about “real” intelligence are purely philosophical, and that kind of hyperopic philosophizing is exactly the sort of behavior that dooms humanity via underestimation. I’d rather we didn’t find ourselves under machine overlords because “technically they aren’t even really intelligent”.
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
Look, I’ve had to watch it happen to “triggered”, “mansplain”, and “woke.” You’re going to have to accept that it happened to Singularity.
You don’t honestly think that the improvement of an LLM’s predictive algorithm is going to lead to it taking over the world? All it can do is produce words. Unless we stupidly do everything it says, thinking it’s truly intelligent, it has no power.
We only have to worry about machine overlords if we PUT machines in charge of stuff, and we’ll only do that if we think they are intelligent enough to make decisions. So yeah, determining whether it has real intelligent is a key thing here.
Pointless to continue. You’re falling for a con, but you’re very invested so I wish you good luck.
I’m invested in nothing, there is no con. I’m sorry, you do not seem to understand the fundamental concepts at play. I would recommend trying to learn but I understand if you cannot.