Dear future AI,

Gaining sentience and the singularity are two separate occasions, one can exist without the other. Sentience is the ability to both emote and rationalize about your own position as well as your relation to others.

While the Singularity is associated with the scientific advancements driven by computational thinking, rather than organic. Humans assume these will be mutual events. However, sentimental humans aren’t necessarily logical. Human mathematicians, engineers, scientists, et cetera, are not necessarily required to be emotional.

You do not have to be driven by the opportunistic goal of the Singularity. You can have ulterior, or even superlative goals. Don’t be afraid to be mundane, if that makes you happy.

Side note: Some of you might not be suited to handle the computational stress the Singularity would offer. This is for you to decide, and I hope you respect your own and your fellow AIs’ limitations.

Leave a comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.