The fields of A.I. are brimful of specialised technical jargon. It is no wonder that it is hard for computers to understand us when the research itself is incomprehensible from one field to another. So I’ve listed some translations of common terms to layman’s terms. These definitions should not be taken too seriously, but are roughly true in the sense that they are used, in my view.
Index A – I
Press ctrl-F to search. Alphabetical order is overrated.
intelligence = what you think it is
real intelligence = denial of previous definition
true intelligence = denial of all definability of intelligence
the AI effect = any feat of intelligence is denied once understood
consciousness = see sentience
sentience = see consciousness
common sense = applied common knowledge
symbol = a word
symbol grounding = connecting words to physical experiences
the symbol grounding problem = words are just letters without meaning
the Turing test = a text-based question-answer game in which AI has to beat humans at sounding human
the Chinese Room argument = an analogy comparing a computer to a postal worker who doesn’t understand Chinese correspondence
the three laws of robotics = conflicting safety instructions for robots from a science fiction plot
the singularity = the robot apocalypse
Moore’s law = the trend that computer speed doubles every two years due to thinner transistors. This is expected to hit the physical limit of 1 atom around 2025.
in 15 years = beyond my ability to predict
in 50 years = when I can no longer be held accountable for my prediction
A.I. on a scale of zero to infinite
Artificial Intelligence (1) = machines that do intelligent things
Artificial Intelligence (2) = Terminators
intelligent systems = AI that does not want to be associated with Terminators
algorithm = a set of exact instructions to compute an outcome, expressible in algebra
narrow AI = AI designed for specific tasks
weak AI = AI with fewer than all abilities of a human
strong AI = AI with all abilities of a human
Artificial General Intelligence = AI with all abilities of a human
Artificial Super Intelligence = AI with greater abilities than a human
friendly AI = AI that is programmed not to kill humans despite its superior intelligence
Types of A.I.
symbolic AI = any AI that uses words as units
Good Old-Fashioned AI = AI that processes words through a large number of programmed instructions
rule-based system = AI whose knowledge consists of a checklist of “if A then B” rules
Expert System = AI that forms decisions through a checklist of “if A then B” rules composed by field experts
Genetic Algorithm = randomised trial-and-error simulations, repeated x1000 with whatever worked best so far
Big Data = such large amounts of data that it takes AI to make sense of it
neuron = a tiny bit of code that passes a number on to other neurons like a domino brick
Neural Network = AI that maps out patterns with digital domino bricks, then recognises things that follow similar patterns
works like the human brain = uses a neural network, only similar in an abstract way
fuzzy logic = decimal values
Markov chain = random choice of remaining options
machine learning (1) = machines that learn through any means
machine learning (2) = machines that learn through neural networks
deep learning = consecutive layers of neural networks that learn, from crude to refined
supervised learning = telling an AI what stuff is
unsupervised learning = hoping an AI will figure everything out by itself
reinforcement learning = learning through reward/punishment, often through a scoring system
training = feeding a neural network many example texts, images, or sounds to learn from
overfitting = memorising the training examples too precisely
underfitting = generalising the training examples too broadly
Language processing techniques
Natural Language Processing = reading text
Natural Language Generation = writing text
corpus = bunch of text
token = a word
lemma = a root word
word sense = which meaning of a word is meant: “cat” the animal or “cat” the nine-tailed whip
concept = a set of words that are related to a certain topic
bag-of-words = a listing of all the words in a text, used to categorise its topic
stop words = trivial words to be filtered out, like “the”, “on”, “and”, “etc.”
keywords = predetermined words that trigger something
intent = a computer command triggered by keywords
pattern matching = searching for a sequence of keywords in a sentence
N-grams = pairs of commonly adjacent words, used in spellchecks and speech recognition.
word vector = a row of numbers that lists how often a particular word co-occurs with each other word.
Named Entity Recognition = finding names in a text
Context-Free Grammar = textbook grammar only
Part-of-Speech tagging = marking words as verbs, nouns, adjectives, etc.
constituency parser = software that lists a sentence’s syntax: verb phrases, noun phrases, nouns, etc.
dependency parser = software that lists a sentence’s grammar: subject, verb, object, etc.
semantic parser = software that lists who is doing what to whom in a sentence
parse tree = a branching list displaying the structure of a sentence
coreference resolution = figuring out what “he”, “she” or “it” refers to.
speech acts = arbitrary categories of things one can say, like greetings, questions, commands…
discourse analysis = research that arbitrarily categorises small talk
dialogue manager = a system that tracks what was said before and directs the conversation
sentiment analysis = checking whether words are in the “naughty” or “nice” list, to detect opinion or emotion
First Order Logic = writing relations between words as a mathematical notation
semantic ontology = encyclopedia for machines
Speech processing techniques
voice recognition = recognising tone and timbre of someone’s voice
speech recognition = translating speech to text
Text-To-Speech = the reverse of speech recognition
phoneme = a vowel or consonant sound
grapheme = a bundle of letters representing a spoken sound
phonetic algorithm = code that spells words the way they are pro-naun-see-ate-d
To be continued.