Professional Documents
Culture Documents
Version (edition), Number (vol. and/or no.), Publisher, Publication Date, Location
(pages, paragraphs and/or URL, DOI or permalink). 2nd container’s title, Other
contributors, Version, Number, Publisher, Publication date, Location, Date of
Access (if applicable).
The Great Hack, Karim Amer & Jehane Noujaim, Netflix.com, JUly 24, 2019
● Title of the site, database, or streaming service that the documentary was found
on (if applicable)
● Name of the production company.
● Publication date.
● City where the production company is based.
Regulation may in fact, be needed. However, there are compelling business reasons that
are inspiring Microsoft and other IT companies to initiate internal and external discussions on
how to manage the increasing ethical complexity of data science, especially AI and deep
learning
.chrome-extension://oemmndcbldboiebfnladdacbdfmadadm/http://datworkshop.org/papers/dat16-final38.pdf
Giffy:
https://media.giphy.com/media/dsKzbZBCEOIw7bAxdV/giphy.gif
privacy and machine intelligence, AI and warfare and criminal uses of AI.
following:
Economics.
AI and warfare.
Loss of control of AI systems. Concerns have been expressed about the possibility of
the loss of control of intelligent machines for over a hundred years. The loss of control
of AI systems has been a recurrent theme in science fiction and has been raised as a
possibility by scientists. Some have speculated that we could one day lose control of AI
systems via the rise of superintelligences that do not act in accordance with human
wishes—and that such powerful systems would threaten humanity. Speculations about
the rise of such uncontrollable machine intelligences have called out different scenarios,
including trajectories that take a slow, insidious course of refinement and faster-paced
evolution of systems toward a powerful intelligence “singularity.” Are such dystopic
outcomes possible? If so, how might these situations arise? What are the paths to these
feared outcomes? What might we do proactively to effectively address or lower the
likelihood of such outcomes, and thus reduce these concerns? What kind of research
would help us to better understand and to address concerns about the rise of a
dangerous superintelligence or the occurrence of an “intelligence explosion”? Concerns
about the loss of control of AI systems should be addressed via study, dialog, and
communication. Anxieties need to be addressed even if they are unwarranted. Studies
could reduce anxieties by showing the infeasibility of concerning outcomes or by
providing guidance on proactive efforts and policies to reduce the likelihood of the
feared outcomes.
Psychology of people and smart machines. We don’t have a good understanding about
how people will feel about and respond to intelligent applications, services, and robots
in their environments and lives. What are the psychological implications of
implementations of intelligence in our various environments? For example, how will our
self-conception—our sense of who we are—change with the rise of machines with
human-like qualities and competencies? What is the evolving nature of relationships
with intelligent software and systems? What is the basis for anxieties about machines
over the centuries, and how might anxieties about smart machines stimulate fear and
discomfort? What is the basis for fears about “superintelligences”? How do popular
media and stories contribute to these fears? We can work to better understand and
study potential dependencies, fears, anxieties, and attractions. Psychological studies
could be undertaken in a proactive manner (e.g., with “Wizard of Oz” methods that
simulate future intelligences) in advance of the availability of such machines.
Neuroscience and AI. Despite many decades of research, we have little genuine
knowledge about how the human brain performs its computational magic. While people
have been using phrases like “brain-inspired” (e.g., with regard to neural network
models, including network models referred to as “deep learning”), we have a
surprisingly poor understanding about the operation of the nervous system, and the
computational methods that it may employ. Allusions to the brain are typically made with
a stretch of the imagination—and with confidence powered largely by intuitions. There is
a rich opportunity for better understanding the impressive capabilities of the brain via
computational models, methods, metaphors, and results. Going in the other direction,
findings about the structure and operation of nervous systems can help AI scientists to
formulate questions and to guide research about computational approaches to thought
and intelligent behavior. Pushing hard on both directions will likely be important, and
there is much to be learned.
AI and philosophy of mind. Philosophers have long grappled with interesting and
compelling questions about the nature of mind, including questions about the nature
and foundations of conscious experience. These discussions often include careful
development of a language to bring discussions to a common ground. Questions have
been asked about whether machines that we build might one day be conscious and find
themselves “aware” and “experiencing” inner or subjective worlds similar to those
experienced by people. Discussions about the computational foundations o