Starting Themes: authority, feeling, assumptions, how people and machines differ (i.e. think); scale of implications/participation; building sustainable future for children/next generation; the implications (job loss, and thus ethics) of AI and future of development of technology); AI and tech for political purpose (command and control); innovation to the sake of innovation (again, missing implications); randomness in a systemized world; IQ and EQ (what is "authentic intelligence" - how do these modalities interact; where is the agency in a systematized future; social media as a catalyst (for good or ill);
- Artificial Intelligence:
- Artificial: man-made vs. sham or false
- Misunderstanding: Information about AI;
- Sentience
- Opposite: Artificial <> Natural
- Pseudo-intelligence (Robert Smith offers as more effective title)
- Challenges to Nick Bostrum's Superintelligence "myth"
- How important is the how vs. the end? Purpose and the value derived through action/activity?
- Is efficiency the most important metric?
- What is our personal responsibility?
- Is curiousity dangerous?
- Social understanding of the AI / tech space? vs. the social consequence of AI/tech?
- Re-imagining authority in the digital space: i.e. participation in the digital space?
- Artificial intelligence ("Sophia")
- Organizational incentives:
- Technology is a by-product of us? How do we embrace/cultivate a humanity that unleashes the technology we want?
- What incentives are there? Do we need to do it differently to incentivize better design?
- If system is geared towards efficiency and calibrate for humanity, how do you design a system to support vulnerability when it is geared towards economic sustainability/survivability?
- The power of human intelligence: is it cognitive? Is it emotional?