Download Research Tools
Interest in the Kinect for Windows Software Development Kit (SDK) beta, released on June 16, 2011, has been strong, and we’re delighted to learn that so many developers and innovators who are experimenting with natural user interface (NUI) applications have taken advantage of the SDK to explore the potential of the Kinect sensor.
In support of our commitment to encourage researchers and enthusiasts in their exploration of the exciting possibilities of the Kinect sensor, we have now released a refreshed version of the SDK. The community has provided us with a lot of good feedback, and this release addresses some of the top items you’ve told us about.
Before summarizing the updates, let’s quickly recap the key features of the Kinect for Windows SDK beta. This non-commercial SDK beta enables human motion tracking, voice recognition, and depth sensing on PCs, enabling developers to create innovative natural user interface applications. The SDK includes drivers and rich APIs for raw sensor streams and natural user interfaces, as well as installation documents and resource materials.
So, what’s in the refresh?
The refresh also includes many improvements to the documentation, including clarifications and the deletion of information pertaining to non-functional components. Also, the SDK samples have been enhanced.
If you’re an academic researcher or an enthusiast who wants to take advantage of the latest developments in natural user interface experimentation, we encourage you to learn more about and download the Kinect for Windows SDK beta refresh. We plan on releasing the next refresh of the Kinect for Windows SDK beta later this year (still with a non-commercial license).
Let us know what you think—as this refresh demonstrates, we’re committed to using your feedback to make the best possible SDK!
—Tony Hey, Corporate Vice President, Microsoft Research Connections
Robotics technology plays an increasingly important role in search-and-rescue missions. Robots are used to explore areas that are deemed too dangerous or difficult for human teams to access. They can, for example, be used to investigate a hazardous material spill or search for disaster survivors. In the case of a disaster, a robot may save the life of not only the victims but also the rescue workers who might otherwise place themselves in harm’s way to search for survivors. Because of the life-saving potential of search-and-rescue robots in emergency situations, researchers are investigating better ways to control the robots in stressful and challenging environments.
Robots that are used for search and rescue are essentially an extension of the human rescue team. Cameras, microphones, and other sensors that are attached to the robot transmit critical information to the rescue team, who typically controls the robot’s movement remotely. Until recently, rescuers who managed a search-and-rescue robot normally had to manipulate complicated joysticks, dials, and switches on a very elaborate controller with multiple electro-mechanical parts. As described in our blog entry last year, the robotics lab at the University of Massachusetts, Lowell (UML), has developed a natural user interface (NUI) controller that promises greater finesse and control of robots during search-and-rescue operations.
Today we are pleased to present a new short video that highlights the accomplishment of this work and gives you an update on its status.
Building the DREAM Controller
The Lowell robotics lab takes a NUI approach for the Dynamically Resizing Ergonomic and Multi-Touch (DREAM) Controller, which has been in development since 2008. Two Microsoft technologies underlie the DREAM Controller: the Microsoft Robotics Developer Studio (used for simulation) and Microsoft Surface (the user interface).
The Microsoft Surface is a coffee-table-sized device with a computer inside and a touch-sensitive interface on top. The Surface allows multiple users to interact with the computer simultaneously by using whole-hand or multiple-finger gestures. These gestures enable rescue teams to control robots with greater dexterity than they could with traditional robotics controllers—and precise control of the robots is critical for search-and-rescue efforts. In addition, the Surface permits more than one robot to be controlled simultaneously—previously not possible with a single controller.
To use the DREAM Controller implemented on the Surface, users simply place their hands on the interface. The DREAM Controller identifies the user’s fingers and thumbs and displays a virtual “joystick” beneath their hand. The user then uses their thumb to manipulate the virtual joystick. There are up to four degrees of freedom (two on each thumb, that is, X-Y on each), enabling the control of four different dimensions.
The Lowell team (Holly Yanco and Mark Micire) is also developing a series of pre-programmed gestures with guidance from expert search-and-rescue volunteers. The goal is to develop code that enables the DREAM Controller to recognize specific gestures that rescue workers make naturally during a search-and-rescue operation, thereby facilitating and accelerating rescue efforts.
The novel NUI approach to robotics that was employed by the Lowell robotics lab in this socially significant application helped the DREAM Controller project win one of eight grants that Microsoft Research offered under our Social Human Robot Interaction Request for Proposals (RFP). The grant award included financial support, a donated Microsoft Surface, and access to the Microsoft Research team.
I think the DREAM Controller project truly shows what a better first response system—using NUI technology—could look like in the very near future. Check out the video!
—Stewart Tansley, Senior Research Program Manager, Microsoft Research Connections
It was barely a year ago that European scientific and industry leaders came together with the goal of developing, testing, and deploying a high-quality, interoperable cloud platform for industry and research. The result was VENUS-C, which stands for Virtual multidisciplinary EnviroNments USing Cloud infrastructures.
Jointly sponsored by the European Commission and a consortium of 14 partners, among them Microsoft Research, VENUS-C was conceived to meet the needs of seven different research and commercial areas: bioinformatics, systems biology, drug discovery, civil engineering, civil protection, civil emergencies, and marine biodiversity. VENUS-C has since developed into a functional, operational platform, and is being used for 15 new pilots that received seed funds after an open call elicited 60 proposals from across Europe.
Conceived to meet the needs of seven research and commercial areas, VENUS-C is now a functional platform that is being used for 15 new pilot programs in various fields.
The success of VENUS-C is hardly surprising, since it offers powerful computing resources, open solutions, and a user-centric focus—all without the upfront costs of expensive IT installations. What’s more, VENUS-C’s massive computing power helps expedite research, speeding the time from hypothesis to result.
Vladimir Sykora, co-founder of Molplex, a small U.K. company working on drug discovery, and recipient of one of the 15 pilot grants, remarks, “Thanks to the VENUS-C platform, we will be able to do in a few weeks molecular computations that would have taken a year to complete on our own servers. This application allows us to quickly estimate the activity in the human body of new chemical compounds.”
VENUS-C also offers the cloud advantages of scalability, providing resources as and when needed. This is enormously valuable in many areas of scientific research when peak computing needs occur sporadically and often unpredictably.
As Costas Papadachos of the Geophysical Laboratory at Aristotle University notes, “Geoscientists involved in the difficult territory of earthquake impact assessment have much to gain from initiatives like VENUS-C. Our involvement offers a prime opportunity to access unprecedented resources, only when and where necessary for earthquake impact estimation and related information dissemination, without worrying how to build and maintain the corresponding infrastructures and operational tools.” Papadachos heads one of the 15 new pilot experiments on the VENUS-C cloud platform.
Cloud computing can provide new approaches to data collection and management, too. Using the VENUS-C platform, Collaboratorio, an Italian-based micro-enterprise that is managing one of the civil engineering scenarios, is collecting data on the performance of new buildings, creating a database of qualitative and statistical information that can be used to find the designs that best fit specific environmental and urban contexts. The cloud will help Collaboratorio’s researchers mine the data to identify trends, perform extrapolation studies, and address common challenges that are related to building and environmental impact.
As VENUS-C embarks on its second year, we look forward to the platform performing even bigger and more complex applications. This is a very precocious one-year-old!
—Fabrizio Gagliardi, Director, Microsoft Research Connections EMEA