Halcyon Dialogue Public Robotics Showcase
In partnership with AAAS, we hosted a public event to celebrate innovations in the fields of AI and robotics. We released the Halcyon Dialogue series report, Shaping Robotics Policy for the 21st Century, and demonstrated the latest in robotics as leading experts discussed our future living with robots and implications of cutting edge developments. And mostly importantly, we met EMIEW3, Microsoft's Magic Wall and experienced emerging technology demonstrations from Johns Hopkins University, Amazon, Carnegie Mellon's Calypso, and more.
Read the 2016-2017 Halcyon Dialogue series report here.
Eager to learn more about the Robots who visited Halcyon? Check out our robotic guests:
Amazon Web Services will demonstrate how to use a fully managed IoT service in the cloud (AWS IoT) as well as an Image Recognition service (AWS Rekognition) to assist in emergency response scenarios. The demonstration will include an IoT connected search device which will be fitted with a camera and LED indicators. Conceptually these devices could be mounted on an unmanned vehicle (drone or land based) or an emergency response vehicle. AWS will have a web interface that allows operators to enter search terms (e.g. fire, person etc.) for the search device to detect. When the search device is activated it will transmit images and location to an image recognition system (running on AWS cloud); service will analyze the incoming images, extract data from it and compare against the list of search terms that were entered into the system. When the system finds a match, operator is notified via the web console to review the image. If the image is a positive match the operator can put the search device into a beacon mode where it will flash its LED’s to notify nearby emergency responders to investigate the area.
Calypso for Cozmo ("Calypso" for short) is a new robot intelligence framwork for the revolutionary Cozmo robot by Anki. Calypso allows anyone age 8-80 to program Cozmo using computer vision, speech recognition, and artificial intelligence algorithms. Calypso was developed by Professor David Touretzky of Carnegie Mellon University. During this live showcase, Dr. Touretzky will demonstrate how Calypso's innovative user interface supports "transparent robot intelligence". #Cozmo #CozmoMoments
EMIEW3 is a humanoid robot with an enhanced degree of autonomy based on EMIEW and EMIEW2. A "remote brain" consisting of control functions deployed on the cloud and a robot monitoring system, forms the robotics IT platform, which enables EMIEW3 to support customer and guidance services. #EMIEW3 #HitachiRobot #Hitachi
This exhibit will highlight robotic systems developed in the Laboratory for Computational Sensing and Robotics at Johns Hopkins University, including new microsurgical robot developed to assist surgeons in minimally invasive applications of otolaryngology, neurosurgery and similar critical fields. The system can eliminate hand tremor and enforce virtual safety barriers to allow surgeons to perform high-stress cases with improved confidence. We will demonstrate a research version of this robot. A clinical version of this robot is being developed by Galen Robotics, Inc.
The Perceptive Pixel (PPI) by Microsoft 55" Touch Device is a touch-sensitive computer monitor capable of detecting and processing a virtually unlimited number of simultaneous on-screen touches. It has 1920 x 1080 resolution, adjustable brightness of up to 400 nits, a contrast ratio of up to 1000:1, and a display area of 47.6 x 26.8 inches. An advanced sensor distinguishes true touch from proximal motions of palms and arms, eliminating mistriggering and false starts. With optical bonding, the PPI by Microsoft 55" Touch Device virtually eliminates parallax issues and exhibits superior brightness and contrast. And it has built-in color temperature settings to accomodate various environments and user preferences. #Surface #SurfaceFam #MagicWall
John Hopkins University Applied Physics Lab (#JHU or #JHUAPL) will present Intelligent Systems that can make decisions under uncertainty and take action when authorized. A legion of trusted, intelligent systems able to autonomously sense, think, decide and act, while interfacing with human teammates on critical applications. JHU/APL has developed Think/Decide algorithms that allow robots to operate “in the wild”. Specifically, their team has focused on enabling a heterogeneous team of robots to:
- perceive and describe novel objects they encounter
- collaborate autonomously on mapping, navigating, and manipulating a challenging environment
- interface with human teammates using natural language
Daniel Turner and Stephen Carter, the co-Founders of TRAXyL and two of Halcyon Incubator's own Cohort 7 fellows, install optical fiber communications on the fly. Using their latest prototype, the FiberTRAXtor, to test their patented installation method, the final product blends in directly with the road surface and allows for digital communications without compromising road surface integrity.