Tag Archives: Possibility Of Ai

How to make that a symbol becomes related to things on which it is not grounded, and a nice introduction to the symbolist/subsymbolist dilemma

Veale, Tony and Al-Najjar, Khalid (2016). Grounded for life: creative symbol-grounding for lexical invention. Connection Science 28(2). DOI: 10.1080/09540091.2015.1130025

One of the challenges of linguistic creativity is to use words in a way that is novel and striking and even whimsical, to convey meanings that remain stubbornly grounded in the very same world of familiar experiences as serves to anchor the most literal and unimaginative language. The challenge remains unmet by systems that merely shuttle or arrange words to achieve novel arrangements without concern as to how those arrangements are to spur the processes of meaning construction in a listener. In this paper we explore a problem of lexical invention that cannot be solved without a model ? explicit or implicit ? of the perceptual grounding of language: the invention of apt new names for colours. To solve this problem here we shall call upon the notion of a linguistic readymade, a phrase that is wrenched from its original context of use to be given new meaning and new resonance in new settings. To ensure that our linguistic readymades ? which owe a great deal to Marcel Duchamp’s notion of found art ? are anchored in a consensus model of perception, we introduce the notion of a lexicalised colour stereotype.

Scientific limitations to the non-scientific idea that super-intelligence will come (for exterminating humans)

Ernest Davis, Ethical guidelines for a superintelligence, Artificial Intelligence, Volume 220, March 2015, Pages 121-124, ISSN 0004-3702, DOI: 10.1016/j.artint.2014.12.003.

Nick Bostrom, in his new book SuperIntelligence, argues that the creation of an artificial intelligence with human-level intelligence will be followed fairly soon by the existence of an almost omnipotent superintelligence, with consequences that may well be disastrous for humanity. He considers that it is therefore a top priority for mankind to figure out how to imbue such a superintelligence with a sense of morality; however, he considers that this task is very difficult. I discuss a number of flaws in his analysis, particularly the viewpoint that implementing ethical behavior is an especially difficult problem in AI research.