• Kinect for Windows Product Blog

    Developing with Kinect for Windows v2 on a Mac

    • 5 Comments

    With the launch of the Kinect for Windows v2 public preview, we want to ensure that developers have access to the SDK so that you can start writing Kinect-based applications. As you may be aware, the Kinect for Windows SDK 2.0 public preview will run only on Windows 8 and Windows 8.1 64-bit systems. If you have a Windows 8 PC that meets the minimum requirements, you’re ready to go.

    For our Macintosh developers, this may be bittersweet news, but we’re here to help. There are two options available for developers who have an Intel-based Mac: (1) install Windows to the Mac’s hard drive, or (2) install Windows to an external USB 3.0 drive. Many Mac users are aware of the first option, but the second is less well known.

    First, you need to ensure that your hardware meets the minimum requirements for Kinect for Windows v2.

    Due to the requirements for full USB 3.0 bandwidth and GPU Shader Model 5 (DirectX 11), virtualization products such as VMWare Fusion, Parallels Desktop, or Oracle VirtualBox are not supported. If you’re not sure what hardware you have, you can find out on these Apple websites:


    Installing Windows on the internal hard drive of your Intel-based Macintosh

    We’re going to focus on getting Windows 8.1 installed, since this is typically the stumbling block. (If you need help installing Visual Studio or other applications on Windows, you can find resources online.)

    Apple has provided a great option called Boot Camp. This tool will download the drivers for Windows, set up bootable media for installation, and guide you through the partitioning process. Please refer to Apple’s website on using this option:


    Alternative to installing Windows on your primary drive

    Boot Camp requires Windows to be installed on your internal hard drive. This might be impractical or impossible for a variety of reasons, including lack of available free space, technical failures during setup, or personal preferences.

    An alternative is to install Windows to an external drive using Windows To Go, a feature of Windows 8 and 8.1 Enterprise. (Learn more about this feature in Windows 8.1 Enterprise.)

    In the section, Hardware considerations for Windows To Go, on Windows To Go: Feature Overview, you can find a list of recommended USB 3.0 drives. These drives have additional security features that you may want to review with your systems administrators, to ensure you are in compliance with your company’s security policies.


    Getting started with Windows To Go

    • You will need the following to proceed:
    • Existing PC with USB 3.0 that has Windows 8/8.1 Enterprise installed (the “technician computer”)
    • USB 3.0 flash or external hard drive
    • Windows 8/8.1 Enterprise installation media (CD or ISO)
    • Windows 8/8.1 product key

    You will need to log in as the administrator. Start the Windows to Go tool, press Win-Q to start the search, and enter Windows To Go:

    press Win-Q to start the search, and enter "Windows To Go"

    Launch the Windows To Go application from the list. From the main application window, you will see a list of the attached drives that you can use with the tool. As shown below, you may be alerted of USB 3.0 drives that are not Windows To Go certified. You can still use the drive but understand that it might not work or could have an impact on performance. If you are using a non-certified USB 3.0 drive, you will have do your own testing to ensure it meets your needs. (Note: while not officially supported by Microsoft, we have used the Western Digital My Passport Ultra 500 GB and 1 TB drives at some of our developer hackathons to get people using Macs up and running with our dev tools on Windows.)

    "Choose the drive you want to use" window

    Select the drive you wish to use and click Next. If you have not already done so, insert the Windows 8.1 Enterprise CD at this time. If you have the .ISO, you can double-click the icon or right-click and select Mount to use it as a virtual drive.

    If you have the .ISO, you can double-click the icon or right-click and select Mount to use it as a virtual drive.

    If you do not see an image in the list, click the Add search location button and browse your system to find the DVD drive or mounted CD partition:

    Browse your system to find the DVD drive or mounted CD partition.

    It should now appear in the list, and you can select it and click Next.

    Select your Windows 8.1 image and click "Next."

    If you need or wish to use BitLocker, you can enable that now. We will Skip this.

    "Set a BitLocker password (optional)" screen 

    The confirmation screen will summarize the selections you have made. This is your last chance to ensure that you are using the correct drive. Please avail yourself of this opportunity, as the Windows To Go installation process will reformat the drive and you will not be able to recover any data that is currently on the drive. Once you have confirmed that you are using the correct drive, click Create to continue.

    "Ready to create your Windows To Go workspace" window

    Once the creation step is complete, you are ready to reboot the system. But first, you’ll need to download the drivers necessary for running Windows on Macintosh hardware from the Apple support page, as, by default, Windows setup does not include these drivers.

    I recommend that you create an Extras folder on your drive and copy the files you’ll need. As shown below, I downloaded and extracted the Boot Camp drivers in this folder, since this will be the first thing I’ll need after logging in for the first time.

    Extracting the Boot Camp drivers from the Extras folder I created.

    Disconnect the hard drive from the Windows computer and connect it to your Mac. Be sure that you are using the USB 3.0 connection if you have both USB 2 and USB 3.0 hardware ports. Once the drive is connected, boot or restart your system while holding down the option key. (Learn more about these startup key shortcuts for Intel-based Macs.)

    Connect the hard drive to your Mac and restart your system while holding down the option key.

    During the initial setup, you will be asked to enter your product key, enter some default settings, and create an account. If your system has to reboot at any time, repeat the previous step to ensure that you return to the USB 3.0 workspace. Once you have successfully logged in for the first time, install the Boot Camp driver and any other applications you wish to use. Then you’ll have a fully operational Windows environment you can use for your Kinect for Windows development.

    Carmine Sirignano
    Developer Support Escalation Engineer
    Kinect for Windows

    Key links

     

  • Kinect for Windows Product Blog

    Media wall promotes engagement at Liberty University

    • 0 Comments

    Ah, the college library: a place of studious reflection and solipsistic quietude—and these days, a whole lot more. On many campuses today, parts of the library have become centers of interactive, multimedia experiences. And perhaps no school has taken this aspect of the modern library further than Liberty University in Lynchburg, Virginia. Serving more than 100,000 students—some 12,000 on campus and more than 90,000 online—Liberty University has established itself as a leader in digital education. Their commitment to using technology to engage and connect with students is front and center in the university’s Jerry Falwell Library, a $50 million, state-of-the-art facility that opened to rave reviews in January 2014.

    Upon entering the library, visitors are greeted by an enormous visual display—a media wall measuring 24 feet by 11 feet, composed of 198 interactive tiles that are controlled by three Kinect for Windows sensors. The wall displays animated visualizations of photos submitted, via social media, by Liberty students and staff. The sensors enable library visitors to use gestures to grab the photos and reveal details about what’s being depicted.

    The media wall thus provides a current snapshot of life at Liberty University—not just on campus but around the world, since most of the school’s students are distance learners. Tim Siegel, systems librarian at Liberty, feels that the wall helps students realize they are part of a bigger picture by “showing the impact that Liberty is having.”

    In order to meet the challenge of providing interaction across the entire width of the wall, Liberty teamed up with InfoStrat, a Kinect for Windows partner located in Washington, D.C., to develop the visualizations and create a custom service that enables the simultaneous use of multiple Kinect for Windows sensors. The sensors are critical to engaging students in this experience and demonstrating Liberty’s technical leadership. Marcy Pride, dean of the library, praises the Kinect feature for the way it draws students in, noting that it “positions us nicely as a twenty-first-century library that uses the technology to engage students, to allow them to have a sense of control, and to give them the opportunity to be creative and innovative.”

    The Falwell Library installation isn’t the first university media wall, but it’s certainly one of the most technologically advanced. The staff surveyed other university installations but wanted to do something bigger and better. When they asked InfoStrat how to make a richer media wall experience, the obvious answer was to make it interactive with Kinect for Windows. The payoff was substantial, as Joshua Blake, technical director of InfoStrat’s Advanced Technology Group, describes. “When students first visit the library, they may not know that the video wall is interactive. As they walk through the space, or they see someone else interacting, it’s a bit of a surprise—this huge video wall responds to them, thanks to the Kinect sensors.”

    As Blake explains, Kinect for Windows is the key to this interactive experience. The arrangement of the three sensors creates a continuous interactive space in front of the entire video wall. These sensors are connected to a single computer via USB extenders, and a Windows Service processes and combines the data from each sensor.

    The media wall visualization is driven by a Unity application, which gets social media and configuration data from a custom web service on a server located elsewhere, and the Kinect interaction data from the local Windows Service through shared memory. The web server ingests the social media posts and images and hosts a management website that allows administrators to moderate posts and remotely configure what is displayed on media wall through the Unity application.

    All that technical wizardry makes the media wall work seamlessly for the visitors who stand in front of it, excitedly waving their arms to move images around and gesturing to access additional information. In the process, they become more deeply connected to the university’s worldwide student body and its many outreach programs. Who knew that the library could be such a potent social force?

    The Kinect for Windows Team

    Key links



  • Kinect for Windows Product Blog

    Hackers push the new Kinect for Windows sensor and preview SDK

    • 0 Comments

    The Kinect for Windows hackathon in Dallas last weekend, July 18–19, was an amazing event, thanks to the 114 developers and designers who willingly gave up their Friday evening and all day Saturday to explore the potential of the new Kinect for Windows v2 sensor and the SDK 2.0 public preview.

    Big thanks to Skip Howard and Computer Visionaries for helping to organize such a great event, to The Dallas Entrepreneur Center for providing the awesome venue, and to local Microsoft evangelists Jason Fox, Nathalie Goh-Livorness, and Anna Lergaard for going above and beyond in collaborating with us in Dallas!
     
    The projects

    My team and I left inspired by the creativity and passion that all of the 15 presented projects displayed. We were absolutely blown away by the three winning efforts. The top prize of US$500 and two Kinect for Windows v2 sensors went to the creators of Super Dueling Golem Dudes, a two-versus-two game in which a player’s gestures control moving platforms that help his or her teammate shoot at their opponent. Earning second-place honors—and $250 and a Kinect for Windows v2 sensor—was the team that developed Kinectergarden, an educational app that lets young children explore their environment and learn new words, objects, and skills. The third-place prize of $100 and a Kinect for Windows v2 sensor was snagged by Put Me in Space, a project that allows users to create a photograph of themselves anywhere in (or out) of this world.

    More than 100 eager hackers put the Kinect for Windows v2 sensor and the SDK 2.0 public preview through its paces.
    More than 100 eager hackers put the Kinect for Windows v2 sensor and the SDK 2.0 public preview
    through its paces.

    And the winners are…

    • Super Dueling Golem Dudes took first place ($500 and Kinect for Windows v2 sensors)

    First-place winners, Super Dueling Golem Dudes

    • Kinectergarten took second place ($250 and a Kinect for Windows v2 sensor)

    Second-place winners, Kinectergarten

    • Put Me in Space took third place ($100 and a Kinect for Windows v2 sensor)

    Third-place winners, Put Me in Space

    The other 12 projects presented on Saturday night were:

    • WaveMeIn, an app that makes passwords passé.  Wave me in, Scotty!
    • Cori App, a digital assistant that controls other applications and connected devices.
    • BorgVision, a utility to help users assimilate into the Borg Kinective.
    • Fly Up, an app that gets you airborne, as your Kinect avatar flies up and off the screen.
    • Sync or Not, an app that maps two people's locations in 3D space and determines if they are moving in sync.
    • STOP!, a potentially lifesaving app that visually detects stop signs and forcefully instructs the driver to stop. 
    • Virtual Skan, an online shopping app that selects the appropriate garment size for the shopper based on a Kinect sensor scan.
    • Talk Charms, an augmentative and alternative communication system that allows nonverbal children to communicate by selecting words and phrases to be played via text-to-speech.
    • Wakie Wakie, another potential lifesaver that prevents drivers from falling asleep at the wheel.
    • Booty Bounce, an app that brings the aerobic power of twerking to your screen, letting you use a variety of twerks to get in shape or just have fun. Miley Cyrus would approve.
    • Clubby VJ, an app that lets you create a cool avatar, whose special effects are perfect for a night of clubbing.
    • Capable Home Kinect, an app that tracks facial movements and uses facial cues to allow quadriplegics to operate devices within the home.

    Upcoming events

    We have more hackathons scheduled this summer, and we’d love to see you at one of them. Just click the link to register.

    Thanks again to all who took the time to join us in Dallas. I got a big jolt of energy from the collected creativity and passion. If you have questions or need help, please use the following resources:

    You can also email me at kinectninja@microsoft.com if you have other questions or just want to say hi.

    I hope to see you at a future event!

    Ben Lower, Developer Community Manager, Kinect for Windows

    Key links



  • Kinect for Windows Product Blog

    The Kinect for Windows v2 sensor and free SDK 2.0 public preview are here

    • 60 Comments

    Today, we began shipping thousands of Kinect for Windows v2 sensors to developers worldwide. And more sensors will leave the warehouse in coming weeks, as we work to fill orders as quickly as possible.

    Additionally, Microsoft publicly released a preview version of the Kinect for Windows SDK 2.0 this morning—meaning that developers everywhere can now take advantage of Kinect’s latest enhancements and improved capabilities. The SDK is free of cost and there are no fees for runtime licenses of commercial applications developed with the SDK.

    The new sensor can track as many as six complete skeletons and 25 joints per person.
    The new sensor can track as many as six complete skeletons and 25 joints per person.

    We will be releasing a final version of the SDK 2.0 in a few months, but with so many of you eagerly awaiting access, we wanted to make the SDK available as early as possible. For those of you who were unable to take part in our developer preview program, now you can roll up your sleeves and start developing. And for anyone else out there who has been waiting—well, the wait is over!

    The new sensor’s key features include:

    • Improved skeletal tracking: The enhanced fidelity of the depth camera, combined with improvements in the software, have led to a number skeletal tracking developments. In addition to now tracking as many as six complete skeletons (compared to two with the original sensor), and tracking 25 joints per person (as compared to 20 with the original sensor), the tracked positions are more anatomically correct and stable—and the range of tracking is broader. This enables and simplifies many scenarios, including more stable avateering, more accurate body position evaluation, crisper interactions, and more bystander involvement in interactive scenarios.
    • Higher depth fidelity: With higher depth fidelity and a significantly improved noise floor, the v2 sensor gives you better 3D visualization, increased ability to see smaller objects and all objects more clearly, and more stable skeletal tracking.
    • 1080p HD video: The color camera captures full, beautiful 1080p video that can be displayed in the same resolution as the viewing screen, allowing for a broad range of powerful scenarios. In addition to improving video communications and video analytics applications, this provides a great input on which to build high-quality, augmented reality scenarios, digital signage, and more.
    • New active infrared capabilities: In addition to allowing the Kinect for Windows v2 sensor to see in the dark, the new infrared (IR) capabilities produce a lighting-independent view, which makes machine learning or computer-vision–based tasks much easier—because you don’t have to account for or model lighting-based variation. And, you can now use IR and color at the same time. We look forward to the many new and innovative uses that the community will develop to use this fundamentally new capability.
    • Wider/expanded field of view: The expanded field of view enables a larger area of a scene to be captured by the camera. As a result, users can be closer to the camera and still in view, and the camera is effective over a larger total area.

    With the ability to track new joints for hand tips and thumbs—as well as improved understanding of the soft connective tissue and body positioning—you get more anatomically correct positions for crisp interactions.
    With the ability to track new joints for hand tips and thumbs—as well as improved understanding of the soft connective tissue and body positioning—you get more anatomically correct positions for crisp interactions.

    In addition to the new sensor’s key features, the Kinect for Windows SDK 2.0 includes:

    • Improved skeletal, hand, and joint orientation: With the ability to track as many as six people and 25 skeletal joints per person—including new joints for hand tips, thumbs, and shoulder center—as well as improved understanding of the soft connective tissue and body positioning—you get more anatomically correct positions for crisp interactions and more accurate avateering. These improved capabilities result in more lifelike avatars and open up new and better scenarios in fitness, wellness, education and training, entertainment, gaming, movies, communications, and more.
    • Support for new development environments: New Unity support provides faster, cost-efficient, and high quality support for cross-platform development, enabling developers to build their apps for the Windows Store using tools they already know.
    • Powerful tooling: Thanks to Kinect Studio’s enhanced recording and playback features, developers can develop on the go, without the need to have a Kinect sensor with them at all times. And Gesture Builder lets developers build their own custom gestures that the system recognizes and uses to write code by using machine learning. These features increase productivity and keep costs down.
    • Advanced face tracking: With significantly increased resolution, applications can capture a face with a 2,000-point mesh that looks more true to life. This means that avatars will look more lifelike.
    • Simultaneous multi-app support: New multi-app support enables more than one application to access a single sensor simultaneously. This means you could have a business intelligence app running at the same time that a training or retail or education experience were running, allowing you to get analytics in real time.

    When the final version of the SDK is available, people will be able to start submitting their apps to the Windows Store, and companies will be able to make their v2 solutions available commercially. We look forward to seeing what everyone does with the new NUI.

    The new SDK 2.0 public preview includes Unity support for faster, cost-efficient, and high quality support for cross-platform development, enabling developers to build their apps for the Windows Store using tools they already know.
    The new SDK 2.0 public preview includes Unity support for faster, cost-efficient, and high quality support for cross-platform development, enabling developers to build their apps for the Windows Store using tools they already know.

    We’ve already shown you what several partners are working on, including Reflexion Health and Freak n’ Genius. Most recently, Walt Disney Studios Motion Pictures have developed an interactive experience to help promote their upcoming movie, Planes 2: Fire & Rescue. One of seven experience kiosks will debut in London at the end of the week in time for school holidays. Disney is confident it will receive an enthusiastic reception from users of all ages, creating an engaging experience associated with the Disney brand and, of course, sparking interest in the movie which releases nationwide from August 8. Read more.

    We will showcase more partner solutions here in coming months, so stay tuned. In the meantime, order your new sensor, download the SDK 2.0 public preview, and start developing your NUI apps. And please join our Microsoft Virtual Academy to learn from our experts and jump start your development.

    The Kinect for Windows Team

    Key links

  • Kinect for Windows Product Blog

    Getting Started with Cinder and Kinect for Windows v2

    • 2 Comments

    On July 15 in a Microsoft Virtual Academy Jump Start session, we will be covering using Kinect for Windows with third-party frameworks. While preparing for the talk, we found that there would be a bit of setup required that we would not have time to cover. So, we are putting together a couple of articles to help as a reference to get your development environment set up and configured.

    If you are not familiar with Cinder or openFrameworks, these are popular open-source graphic frameworks focusing on native C++ development. Both are extensible through plug-ins called Blocks and Add-ons respectively and have extensive libraries for handling 2D and 3D graphics applications. The following articles will provide resources and steps needed to start developing a Windows application by using Kinect for Windows.

    Carmine Sirignano
    Developer Support Engineer, Kinect for Windows

    Key links

  • Kinect for Windows Product Blog

    Kinect with soccer and suds

    • 0 Comments

    What’s better than watching your team compete in the World Cup? Perhaps enjoying the game with a refreshing beer, especially if you purchased the brew at a nice discount. Thanks to a Kinect for Windows application, Argentine fans were able to do just that.

    The Kinect-based display rewarded Argentine soccer fans with discounts on beer.
    The Kinect-based display rewarded Argentine soccer fans with discounts on beer.

    The application was created by Kimetric, an Argentine company that specializes in real-time customer analytics. It involved a special World Cup display, which featured a video of renowned Argentine footballer Oscar Ruggeri, placed in the beer aisle at several supermarkets in Buenos Aires. The Kinect for Windows sensor’s camera detected when customers engaged with the display, and at that point, Ruggeri’s video persona asked them if they’re old enough to drink. If the customers answered yes, the system scanned them to see if they did indeed appear to be over the legal drinking age. Then the sensor looked to see if they were wearing an Argentine soccer jersey. The Kinect for Windows sensor could identify whether or not a customer was clad in Argentine soccer attire based on a machine-learning algorithm. More than 50,000 images of different Argentine soccer jerseys and other apparel were used to train the system.

    So, what happened if the customer was wearing an Argentine jersey? He or she was rewarded with a discount coupon for the purchase of Quilmes Cristal, a popular Argentine beer. If the Kinect for Windows sensor detected just one jersey-clad customer, the discount was 10 percent. If the sensor detected two such customers, the discount rose to 15 percent, and if there were three or more customers posing in Argentine soccer apparel, the discount jumped to 25 percent. Bringing your soccer-crazy friends with you on a beer run really paid off.

    If the customer wasn’t wearing an appropriate jersey, he or she still got a shot at scoring a beer bargain. In those cases, Ruggeri asked a couple of questions about Argentine soccer, and customers who answered correctly were rewarded with a 10 percent discount coupon.

    The display was in stores from early May until the end of June 2014, during which time tens of thousands of lucky customers scored a beer discount. One lucky, randomly chosen customer even received an all-expenses-paid trip to Brazil to attend one of the Argentine team’s matches. Of course, another winner was Quilmes, which saw a healthy increase in beer sales during the promotion.

    All and all, a win for everyone: soccer fans, beer lovers, Quilmes, and Kimetric.

    The Kinect for Windows Team

    Key links

  • Kinect for Windows Product Blog

    Hackers put Kinect for Windows v2 through its paces

    • 1 Comments

    The Kinect for Windows team recently traveled to New York City for a terrific hackathon. In partnership with NUI Central, we had the pleasure of sponsoring a 27-hour event, June 21–22, focused on creating innovative applications and experiences using the forthcoming Kinect for Windows v2 sensor and SDK 2.0.

    The event drew more than 100 participants, from as far away as the Czech Republic and Dubai. You will find a summary of the teams and projects below; we were blown away by what we saw. We were also impressed by the tremendous interest in the experimental near-field firmware that we brought. This firmware turns the Kinect for Windows v2 sensor into a near-field device that can be used at close range (from 10cm to 1m).

    Lots of participants told us they wanted to get their hands on this near-field firmware. Others said how pleased they were just to have access to our engineering team. In fact, there was so much interest and energy at this event that we have set up some additional ones this summer. We would love to see you at one of our upcoming in-person events, or to have you join us online for our virtual training event on Tuesday, July 15. Please see the in-person event information below.


    Upcoming in-person events

    All of these events will be attended by members of the Kinect for Windows engineering team and will feature prizes, food, and, best of all, access to Kinect for Windows v2 sensors (including experimental near-field).

    July 18–19, 2014: Dallas Entrepreneur Center (Dallas, TX, United States)

    Computer Visionaries is hosting; we are sponsoring. It’s going to be so hot outside that you’ll want to be hunkered down inside at our event.


    July 26–27, 2014: Microsoft Headquarters (Redmond, WA, United States)

    This event will be here on our main campus. We will have our engineers out in full force to support this event.

    August 8–10, 2014: Kitchener Studio Space (Kitchener, ON, Canada)

    We are sponsoring Deep Realities and hosting a weekend hackathon in the Waterloo Region.


    New York City hackathon summary

     New York City Hackathon participants
    New York City hackathon participants


    As noted earlier, this event was hosted by our partner, NUI Central, in Manhattan (New York). The top three teams were:

    First place: lightspeed. Their application, called K4B, used a Kinect for Windows v2 sensor to scan large ceiling spaces in order to map HVAC and electrical details so that renovators can get accurate, detailed as-built measurements.

    First-place winners, team lightspeed
    First-place winners, team lightspeed


    Second place:
    adabiits. Their application, Thing T. Thing, included a robotic hand that waves when someone walks by and can be controlled by using finger tracking.

    Second-place winners, team adabiits
    Second-place winners, team adabiits


    Third place:
    Body Labs. Their application, ScanAdHoc, combined multiple Kinect for Windows v2 sensors wirelessly over WebSockets, enabling accurate 3D body scans to be used for fitting clothes.

    Third-place winners, Body Labs
    Third-place winners, Body Labs


    Other teams that presented projects:

    • Augmented Travel presented “Augmented Travel,” which allows you to explore an area virtually, by using your body as the controller.
    • Cornell Tech presented “Body DJ,” which enables you to “rock out” by using body gestures.
    • Critical Mass presented “Sweeper,” which uses Kinect Ripple to raise awareness about the deadly nature of landmines. The team posted a video.
    • Gunters presented “Blox,” which combines the v2 sensor and Oculus Rift to create a personal world using drag and drop.
    • Landroids presented “Obstacle Avoidance Robot,” which used Kinect for Window’s depth sensing capabilities to aid navigation for a Lego Mindstorm robot.
    • Leftover presented “Touchless Controller,” which enables you to use gestures to control any Windows application or to scroll and navigate in a web browser or Visual Studio.
    • Sandwich Dance Party presented “Hue Light Controller,” which enables you to point at a physical light in the room to turn it on or off, and then to use a hand gesture to change the color of the light.
    • SLIP SLAP presented “Around the Cluck,” a game about gold and a guardian chicken, created in Unity 3D. Players must move around the room and carefully collect eggs, so as not to awaken the chicken.
    • SoundPound presented “SoundPound,” a human drum that functions as roadside sobriety test.
    • SpiceDoctor presented “Kinect Virtual Doctor,” a Windows 8 application that uses the Kinect for Windows v2 sensor to monitor your heart rate in real time.
    • Team Fantabulous presented “Power Doge,” a cutting-edge media browser in which novel hand gestures create a unique photo viewing experience.
    • Trail Makers presented “Immersive Trail,” a twenty-first–century approach to the Trail Making Test, a neuro-psychological assessment of visual attention and the ability to switch tasks.
    • Tropicana Pure Premium presented “Finger Tracking Keyboard,” which enables you to shift keys by moving your hand.

    Kinect for Windows MVP András Velvárt helps team adabiits.
    Kinect for Windows MVP András Velvárt helps team adabiits
    .

     

     Hannes Hofman and Chris Schaller from Metrilus brought their finger-tracking library to the event.
    Hannes Hofman and Chris Schaller from Metrilus brought their finger-tracking library to the event.


    Congratulations to all the participants, whose creativity highlighted the immense possibilities in the Kinect for Windows v2 sensor and SDK. If you haven’t pre-ordered the v2 sensor yet, there is still time, but don’t wait too long.

    Ben Lower, Kinect for Windows developer community manager
    Contact Ben on Twitter at @benlower

    Key links

     

  • Kinect for Windows Product Blog

    Keeping watch in the night

    • 0 Comments

    The following blog post was guest authored by Ana Isabel Zorrilla, project manager at EIC BBK-Dravet Syndrome Foundation, a Spanish nonprofit organization dedicated to the treatment and cure of Dravet syndrome and related disorders.

    Imagine what it’s like to go for weeks on end without a decent night’s sleep. For Julian Isla of Madrid, Spain, this scenario requires no imagining. He has lived it, spending night after sleepless night lying awake, listening for the telltale signs that his child is having a seizure. You see, Julian’s son, Sergio, suffers from Dravet syndrome, a severe form of epilepsy. Children with Dravet syndrome experience frequent seizures; from an average of one crisis per week in the mildest cases, to one or more a day—or even multiple seizures per hour—in the most severe instances. These attacks occur more frequently while a child sleeps, so parents often struggle to stay awake seven nights a week, prepared to give their child emergency medical treatment in the event of a prolonged seizure.

    When he was born, Sergio was a healthy baby. The seizures began when he was four months old, and he was diagnosed as having Dravet syndrome after experiencing several long-lasting seizures and multiple admissions to the pediatric intensive care unit. After receiving his son’s diagnosis five years ago, Julian connected with other Dravet families, and together they explored the idea of creating an organization to promote research on the condition. They got in touch with the Dravet Syndrome Foundation (DSF) in the United States, and out of that connection, the Spanish Delegation of the DSF was born. Julian serves as the executive chairman of the Spanish Delegation, which has nine employees, including me. Our organization is involved in multiple research projects, including a search for new drugs that can treat Dravet syndrome.

    Technology has been pivotal for our group; this year, for example, we conducted the world’s first genetic tests for Dravet syndrome, using state-of-the-art technologies. Today, children with Dravet syndrome are being diagnosed by a new generation of genetic tests running on Microsoft Azure that the Spanish Delegation pioneered. Our group’s technological bent is hardly surprising given Julian’s background: with a degree in computer science and some 20 years of professional experience in the IT sector, he currently works for Microsoft in Madrid, where he manages a team of software consultants.

    As soon as Julian heard about Kinect for Windows, he began exploring the use of the technology to help Dravet families. He was aware of the technology’s potential for medical applications, thanks to interactions with colleagues in the Microsoft Madrid offices. About the same time, the Spanish Delegation created EIC BBK, a development center focused on e-health applications. Julian proposed that the center investigate the use of Kinect for Windows to monitor Dravet children while they slept. With the Kinect sensor serving as a sentinel, Julian thought the beleaguered parents might get some much needed sleep themselves.

    The use of monitors to detect seizures is not groundbreaking: there have been multiple studies and projects on seizure monitor systems. What is new and exciting is the Kinect sensor’s exquisite sensitivity, the result of its multisensory inputs. “With a color camera, an infrared detector, and an array of microphones, the Kinect sensor can detect physical movement and acoustical changes with tremendous accuracy,” says Julian. He adds, “The affordability of the Kinect sensor is another huge advantage.”

    By the end of 2013, EIC BBK had started the “Night Seizure Monitor” project, a research initiative that uses Kinect for Windows. This project’s aim is to track the child’s movements while sleeping. When the Kinect sensor detects movements that follow a seizure pattern, an alarm warns parents that their child might be having a seizure. This solution provides dual benefits: when a seizure is detected, the monitor system ensures that the child gets medication right away to reduce the length and intensity of the episode. And when no seizures occur, the monitor enhances family’s quality of life, because parents are able to enjoy a restful sleep.

     At the outset, developers programmed the Kinect sensor to be able to detect the movements of a child even if he or she was in a darkened room and lying under a blanket or comforter.

    At the outset, developers programmed the Kinect sensor to be able to detect the movements of a child even if he or she was in a darkened room and lying under a blanket or comforter, above. Then, they added the ability to spot seizures that begin with abrupt movements or loud vocalizations, below.

     Then, the developers added the ability to spot seizures that begin with abrupt movements or loud vocalizations.

    Using data collected by its color camera and depth sensor, Kinect for Windows detects seizures by comparing changes in the child’s body position between two sequential frames. If these changes are frequent, the seizure alarm sounds. In addition, the sensor’s microphone plays a role in recognizing the seizures, as the system has been programmed to respond to the shouts that typically accompany the onset of a Dravet seizure. The Kinect-based solution can process this sound and calculate the child’s location in the room.

    As participants in the Kinect for Windows Developer Preview program, our developers here at EIC BBK have been testing a preview version of the Kinect for Windows v2 sensor and SDK since December 2013, exploring the technology’s potential to improve upon the existing Night Seizure Monitor research. They are especially pleased with the v2 sensor’s infrared capabilities—which provide an even higher quality image—and its wider field of vision and greater depth range. These enhanced features should make the Night Seizure Monitor even more valuable.

    Moreover, the developers are eagerly awaiting the general availability of the Kinect for Windows v2 sensor and SDK, which promise enhanced discrimination of facial expressions. The developers believe the enhanced face tracking capability will help the monitor detect those seizures that do not present limb shaking but rather are manifested by movements of the eyes and mouth.

    The Night Seizure Monitor initiative is a great example of how needs can promote creativity. Julian had a problem at home, and rather than accepting it as unavoidable, he decided to seek out a solution.

    It also shows the power of teamwork: Julian has received enormous support from colleagues at Microsoft; right now a dozen of them are helping our organization as volunteers.

    Finally, it demonstrates how technology can empower people. Julian sums up the experience eloquently, observing that “When you have a child with special needs, everything seems filled with problems. You feel impotent. I’ve had the privilege of using technology for a project that will, we believe, improve the lives of many young patients and provide a sense of control to their families. I feel proud to work for a company whose technology can make such a difference in people’s lives.”

    While the Night Seizure Monitor is still a development project, we hope to have a fully functional prototype available for testing with Dravet Foundation families by the end of 2014. After that, our goal is to make the monitor available to Dravet families around the world. But as wonderful as this development will be, it is but a way station in the Dravet Syndrome Foundation’s ultimate mission: to find a cure for this disorder. Despite the difficulty of this quest, DSF supporters, volunteers, and workers are laboring tirelessly to achieve it. We call it our “moonshot,” taking inspiration from US President John Kennedy’s audacious mission to send a manned mission to the moon. Our moonshot represents the dream of parents who will never give up.

    Ana Isabel Zorrilla, project manager,
     EIC-BBK– Dravet Syndrome Foundation

    Key links

             

     

  • Kinect for Windows Product Blog

    Pre-order your Kinect for Windows v2 sensor starting today

    • 32 Comments

    At BUILD in April, we told the world that the Kinect for Windows v2 sensor and SDK would be coming this summer, and with them, the ability for developers to start creating Windows Store apps with Kinect for the first time. Well here in Redmond, Washington, it’s not summer yet. But today we are pleased to announce that developers can pre-order the Kinect for Windows v2 sensor. Developers who take advantage of this pre-order option will be able to start building solutions ahead of the general public.

    Sensors purchased during the pre-order phase will be shipped in July, at which time we will also release a public beta of our software development kit (SDK). All of this will happen a few months ahead of general availability of sensors and the SDK, giving pre-order customers a head start on using the v2 sensor’s new and improved features, including increased depth-sensing capabilities, full 1080p video, improved skeletal tracking, and enhanced infrared technology.

    Kinect for Windows v2 sensor

    Thousands of developers wanted to take part in our Developer Preview program but were unable to do so—in fact, we’re still receiving requests from all around the world. So for these and other developers who are eager to start using the Kinect for Windows v2, the pre-order option offers access to the new sensor ahead of general availability. Bear in mind, however, that we have limited quantities of pre-order sensors, so order while supplies last.

    The v2 sensors will also be shipped in July to those who participated in the Developer Preview program. For these early adopters, it’s been an amazing six months: we’ve seen more stunning designs, promising prototypes, and early apps than we can count—from finger tracking to touch-free controls for assembly line workers to tools for monitoring the environment. At BUILD, we showed you what Reflexion Health and Freak’n Genius were able to achieve with the v2 sensor in just a matter of weeks. And in July, when the sensor and SDK are more broadly available, we can only imagine what’s next.

    Kinect for Windows will continue to feature more innovative uses of the v2 technology on this blog in the coming months. As Microsoft Corporate Vice President and Chief Evangelist Steven Guggenheimer notes, “I love what the Kinect sensor and SDK can do. Getting the v2 sensor into the hands of more developers and getting the SDK more widely available is the next step.”

    We are committed to a future where humans and technology can interact more seamlessly—in the living room, on their PCs, and beyond.

    —The Kinect for Windows Team

    Key links


  • Kinect for Windows Product Blog

    Kinect-powered stroke rehab system gets FDA clearance

    • 0 Comments

    Back in January, we featured a story about Jintronix and its innovative rehabilitation system that uses Kinect for Windows to help stroke victims recover physical functions. The origins of that story date back to April 2012, when Jintronix, then based out of Montreal, became one of 11 startups selected from more than 500 applicants worldwide to take part in the Kinect Accelerator program. That program was powered by Techstars—one of the world’s foremost technology accelerators—in close collaboration with Microsoft.

    Now, as Dr. Bill Crounse, senior director of Worldwide Health at Microsoft reports in his blog, Jintronix has received 510(k) clearance from the US Food and Drug Administration (FDA) for its rehab system. This marks an important milestone for Jintronix. “We’re very excited about receiving FDA clearance, which paves the way for Jintronix to help in the rehabilitation of countless stroke victims,” said CEO Shawn Errunza.

    The Kinect for Windows Team

    Key links

Page 2 of 11 (106 items) 12345»