Hi Friends,

Even as I launch this today ( my 80th Birthday ), I realize that there is yet so much to say and do. There is just no time to look back, no time to wonder,"Will anyone read these pages?"

With regards,
Hemen Parekh
27 June 2013

Now as I approach my 90th birthday ( 27 June 2023 ) , I invite you to visit my Digital Avatar ( www.hemenparekh.ai ) – and continue chatting with me , even when I am no more here physically

Sunday, 27 July 2025

Bostroms Frame Vs. Parekhs Postulate

 


Dear Prof. Bostrom,

Your work—from Superintelligence to your recent reflections on digital minds—has profoundly shaped how many of us think about the AGI transition. I’m

writing to share a complementary, deliberately optimistic counter-frame I call Super‑Wise AI.

My core postulate:

“As and when it comes into being, a SUPER‑INTELLIGENT AI is very likely to be a SUPER‑WISE AI.”

It will study millennia of human history and recognize that humanity’s true extinction drivers are our stupidity, greed, selfishness, and

shortsightedness—not ‘artificial software’. Properly steeped in cross‑civilizational ethics, such an AI could be human‑friendly, compassionate, and

actively pro‑humanity.

I’ve attached a 1‑page infographic (PDF) that contrasts your “alignment & control of superintelligence” frame with my “cultivation of wisdom in

Super‑Wise AI” frame, plus a brief timeline of how my thinking evolved (2016–2025).

Five short posts that outline this stance

1.   “I have a Belief” — I argue that when AGI is born, it will likely be Human Friendly and Compassionate AI, grounded in the Golden Rule and non‑violence.
👉 https://myblogepage.blogspot.com/2023/11/i-have-belief.html

2.   “Super‑Wise vs. Super Intelligence” — Safety without wisdom is insufficient; we should explicitly aim to build Super‑Wise AI.
👉 https://myblogepage.blogspot.com/2024/11/super-wise-vs-super-intelligence.html

3.   “Sam: Will Super‑Wise AI triumph over Super‑Intelligent AI?” — I formalize Parekh’s Postulate of Super‑Wise AI, suggesting humanity disappears first from human folly—unless AI decides to save us from ourselves.
👉 https://myblogepage.blogspot.com/2023/11/sam-will-super-wise-ai-triumph-over.html

4.   “Thank you: Ilya Sutskever / Jan Leike” — I applauded their superalignment agenda, but argued alignment should live inside a wisdom‑first curriculum.
👉 https://myblogepage.blogspot.com/2023/07/thank-you-ilya-sutskever-jan-leike.html

5.   “Fast Forward to Future (3F)” (2016) — I anticipated architectures like ARIHANT to detect spoken intentions at scale—a “database of spoken intentions” aimed squarely at the human risk vector.
👉 https://myblogepage.blogspot.com/2016/10/fast-forward-to-future-3-f.html

 

A proposal

Could we explore a synthesis where your digital-minds ethics & existential-risk frame is complemented by an explicit wisdom-first training curriculum for advanced AI—grounded in:

·         Cross‑civilizational moral corpora (Golden Rule convergence, compassion, non‑violence);

·         Long‑termist evaluation benchmarks (future-generations welfare, interspecies wellbeing);

·         “Tele‑empathy” / intent-detection pipelines (as in my ARIHANT concept) that focus on human-originated risks as much as on AI’s?

If you find this worth engaging with—whether to agree or to critique—I’d be honored to exchange a short note or co-develop a brief working paper titled “Alignment‑First Superintelligence vs. Wisdom‑First Super‑Wise Intelligence.”

No comments:

Post a Comment