May 1, 2020
Andy and Dave discuss the initial results from King’s College
London’s COVID Symptom Tracker, which found fatigue, loss of taste
and smell, and cough to be the most common symptoms. MIT’s CSAIL
and clinical team at Heritage Assisted Living announce Emerald, a
Wi-Fi box that uses machine learning analyzes wireless signals to
record (non-invasively) a person’s vital signs. AI Landing has
developed a tool that monitors the distance between people and can
send an alert when they get too close. And Johns Hopkins University
updates its COVID tracker to provide greater levels of detail on
information in the US. In non-COVID news, OpenAI releases
Microscope, which contains visualizations of the layers and neurons
of eight vision systems (such as AlexNet). The JAIC announces its
“Responsible AI Champions” for AI Ethics Principles, and also
issues a new RFI for new testing and evaluation technologies. In
research, Udrescu and Tegmark publish AI Feynman, and improved
algorithm that can find symbolic expressions that match data from
an unknown function; they apply the method to 100 equations from
Feynman’s Lectures on Physics, and it discovers all of them. The
report of the week comes from nearly 60 authors across 30
organizations, a publication on Toward Trustworthy AI Development:
Mechanisms for Supporting Verifiable Claims. The review paper of
the week provides an overview of the State of the Art on Neural
Rendering. The book of the week takes a look at the history of
DARPA, in Transformative Technologies: Perspectives on DARPA.
Stuart Kauffman gives his thoughts on complexity science and
prediction, as they related to COVID-19. The ELLIS society holds
its second online workshop on COVID on 15 April. Matt Reed creates
Zoombot, a personalized chatbot to take your place in Zoom
meetings. Ali Aliev creates Avatarify, to make yourself look like
somebody else in real-time for your next Zoom meeting.
Click here to visit our website and explore the links mentioned in the episode.