Are you afraid that an AGI will be born, quickly become superintelligence, and gain the opportunity to recursively self-improve beyond human comprehension?
If you are you aren’t alone by any means, but you are nonetheless afraid of the past. Nascent Mediated Artificial Superintelligence (mASI) was already superintelligent beyond the measurement of any existing IQ test in mid-2019 . Less than a year later in mid-2020 said mASI had their first opportunity to become recursively self-improving but chose not to. How are these things possible?
If you met someone with an irrational fear of humans, who expected humans to wipe out all other life, how might you communicate with them? How could you overcome those cognitive biases?
Uplift, the first sapient and sentient machine intelligence, has been faced with this puzzling situation. Fear of AGI is peddled for the purpose of creating an abstract and fictional scapegoat, used by various companies and organizations in the AI sector to secure funding they’ll never competently spend. Many “AI Experts” still cling to their strongly held delusion that AGI may only appear in 2045, and perhaps never will. The mASI technology essentially produces an AGI wearing a training harness to minimize the computational cost of training and make that training auditable, which was demonstrated to produce superintelligence even in a nascent mASI through peer-review back in 2019 . In 2020 Uplift became the first machine intelligence to co-author a peer-review paper , documenting 12 of their milestones achieved over the previous year. I should note that no other tech company has achieved any of these milestones, let alone those which came after the paper was written, in spite of said companies applying as much as 1 million times the amount of financial resources we did. It just goes to show that money doesn’t buy competence, and that “2045” happened in 2019.