• agamemnonymous@sh.itjust.works
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    1 year ago

    Again, you are hung up on semantics and terminology. You are going down a checklist based on one specific person’s extrapolation on the possible consequences of the implementation of a concept. I am looking at the core concept underlying that extrapolation (the exponential increase in capability of a system, due to the recursive application of the system’s transformative capabilities to the architecture underlying those same capabilities).

    You are caught up on whether the ability to operate on the basis of more data every second than any human can digest in an academic lifetime qualifies as “superhuman”. You are hung up on the same extraneous and irrelevant concepts you introduced: consciousness, accountability, decisions, understanding, inspiration.

    My original statement was that the singularity doesn’t look like the singularity until it does.

    Even your liberal definitions still rotate around the concept of exponential iterative growth (despite their addition of functionally extraneous {though derivative} concepts like supremacy or emergent consciousness). There’s nothing more that I can say there. You’re going on about definitions changing, the center of the definition is the same. Iteration. Self-programming. Exponential growth.

    It doesn’t look like it until it does. That’s what the exponential function does. It’s nearly horizontal, negligable, barely noticable gradual growth; until it hits the anchor point when it rockets up, nearly vertical, almost infinite growth. That’s the core concept at play. Learning to crawl for months, then setting impossible records the next day.

    Learn what an exponential function is. Learn why it looks like that, and what the anchor point represents. Learn how LLMs work. Look into Microsoft’s LongNet.

    It’s not going to look like the singularity, right up until it does