<nodes> <node id="210251">  <title><![CDATA[Robots Able to Reach through Clutter with Whole-Arm Tactile Sensing]]></title>  <uid>27462</uid>  <body><![CDATA[<p>Whether reaching for a book out of a cluttered cabinet or pruning a bush in the backyard, a person’s arm frequently makes contact with objects during everyday tasks. Animals do it too, when foraging for food, for example.</p><p>Much in the same way, robots are now able to intelligently maneuver within clutter, gently making contact with objects while accomplishing a task. This new control method has wide applications, ranging from robots for search-and-rescue operations to assistive robotics for people with disabilities.</p><p>“Up until now, the dominant strategies for robot manipulation have discouraged contact between the robot’s arm and the world,” said Charlie Kemp, lead researcher and associate professor in the <a href="http://www.bme.gatech.edu/">Coulter Department of Biomedical Engineering at Georgia Tech and Emory University</a>. “Instead of avoiding contact, our approach enables the arm to make contact with objects, people and the rest of the robot while keeping forces low.”</p><p>Kemp, director of Georgia Tech’s Healthcare Robotics Lab, along with his graduate students and researchers at <a href="http://mekabot.com/">Meka Robotics</a>, has&nbsp;developed a control method that works in tandem with compliant robotic joints and whole-arm tactile sensing. This technology keeps the robot’s arm flexible and gives the robot a sense of touch across its entire arm.</p><p>With their control method, Kemp’s robots have performed numerous tasks, such as reaching through dense artificial foliage and a cinder block representative of environments that search-and-rescue robots can encounter.</p><p>A publication describing the research, <a href="http://intl-ijr.sagepub.com/content/32/4/458">“Reaching in Clutter with Whole-arm Tactile Sensing</a>,” appears in this month’s edition of the <em>International Journal of Robotics Research</em>.&nbsp;</p><p>Kemp's lab also has promising results that could impact the future of assistive robotics. They have developed tactile sensors made out of stretchable fabric that covers the entire arm of a robot. In a preliminary trial with the new control method and sensors, Henry Evans, a person with quadriplegia, used the robot to perform tasks for himself. He was able to pull a blanket over himself and grab a cloth to wipe his face, all while he was in bed at his home.</p><p>This trial was conducted as part of the Robots for Humanity project with Willow Garage. In order to ensure safety, researchers from Kemp’s lab closely monitored the activities. This research has been accepted and will be presented at the <a href="http://depts.washington.edu/uwconf/icorr2013/">International Conference on Rehabilitation Robotics</a> in June.&nbsp; </p><p>“I think it’s a good safety feature because it hardly presses against me even when I tell it to,” Evans said after the trial. “It really feels safe to be close to the robot.”</p><p>Evans was also impressed by how the robot’s arm “just wriggles around obstacles.”</p><p>Kemp’s research team has also released the designs and code for the sensors and controller as <a href="http://www.hsi.gatech.edu/hrl/project_open_source_whole_arm_tactile_sensing.shtml">open source hardware and software</a> so that researchers and hobbyists can build on the work.</p><p>The research is part of an ongoing effort to create a new foundation for robotics, where contact between the robot’s arm and the world is encouraged.</p><p>“Our belief is that this approach is the way of the future for robots,” said Kemp, who is also a member of Georgia Tech’s <a href="http://robotics.gatech.edu/">Center for Robotics and Intelligent Machines</a>. “It is going to allow robots to better operate in our homes, our workplaces and other complex environments.”</p><p><em>This research is funded by the DARPA Maximum Mobility and Manipulation (M3) Contract W911NF-11-1- 603. The assistive technology research is also funded in part by NSF CAREER award IIS-1150157, NSF grant CNS-0958545, an NSF GRFP and Willow Garage.</em></p><p><strong>CITATIONS</strong>: Advait Jain, Marc D Killpack, Aaron Edsinger, and Charles C Kemp. Reaching in Clutter with Whole-arm Tactile Sensing. The International Journal of Robotics Research, April 2013, 32: 458-482, doi:10.1177/0278364912471865</p><p>Phillip M. Grice, Marc D. Killpack, Advait Jain, Sarvagya Vaish, Jeffrey Hawke, and Charles C. Kemp. Whole-arm Tactile Sensing for Beneficial and Acceptable Contact During Robotic Assistance. Accepted to the 13th International Conference on Rehabilitation Robotics (ICORR), June 2013.&nbsp;</p>]]></body>  <author>Liz Klipp</author>  <status>1</status>  <created>1367250499</created>  <gmt_created>2013-04-29 15:48:19</gmt_created>  <changed>1475896448</changed>  <gmt_changed>2016-10-08 03:14:08</gmt_changed>  <promote>0</promote>  <sticky>0</sticky>  <teaser><![CDATA[Robots are now able to intelligently maneuver within clutter, gently making contact with objects while accomplishing a task, thanks to technology developed by Dr. Charlie Kemp and the Healthcare Robotics Lab.]]></teaser>  <type>news</type>  <sentence><![CDATA[Robots are now able to intelligently maneuver within clutter, gently making contact with objects while accomplishing a task, thanks to technology developed by Dr. Charlie Kemp and the Healthcare Robotics Lab.]]></sentence>  <summary><![CDATA[<p>Robots are now able to intelligently maneuver within clutter, gently making contact with objects while accomplishing a task, thanks to technology developed by Dr. Charlie Kemp and the Healthcare Robotics Lab.</p><p>&nbsp;</p><p>&nbsp;</p>]]></summary>  <dateline>2013-04-29T00:00:00-04:00</dateline>  <iso_dateline>2013-04-29T00:00:00-04:00</iso_dateline>  <gmt_dateline>2013-04-29 00:00:00</gmt_dateline>  <subtitle>    <![CDATA[]]>  </subtitle>  <sidebar><![CDATA[]]></sidebar>  <email><![CDATA[klipp@gatech.edu]]></email>  <location></location>  <contact><![CDATA[<p><strong>Georgia Tech Media Relations</strong><br />Laura Diamond<br /><a href="mailto:laura.diamond@comm.gatech.edu">laura.diamond@comm.gatech.edu</a><br />404-894-6016<br />Jason Maderer<br /><a href="mailto:maderer@gatech.edu">maderer@gatech.edu</a><br />404-660-2926</p>]]></contact>  <boilerplate></boilerplate>  <boilerplate_text><![CDATA[]]></boilerplate_text>  <media>          <item>210121</item>          <item>210131</item>          <item>210141</item>          <item>210151</item>      </media>  <hg_media>          <item>          <nid>210121</nid>          <type>image</type>          <title><![CDATA[Robots Reaching Through Clutter]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[kemp_robot3.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/kemp_robot3_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/kemp_robot3_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/kemp_robot3_0.jpg?itok=-gD53m6I]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Robots Reaching Through Clutter]]></image_alt>                    <created>1449180018</created>          <gmt_created>2015-12-03 22:00:18</gmt_created>          <changed>1475894869</changed>          <gmt_changed>2016-10-08 02:47:49</gmt_changed>      </item>          <item>          <nid>210131</nid>          <type>image</type>          <title><![CDATA[Robots Reaching Through Clutter - 1]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[kemp_robot4.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/kemp_robot4_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/kemp_robot4_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/kemp_robot4_0.jpg?itok=0posZeHL]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Robots Reaching Through Clutter - 1]]></image_alt>                    <created>1449180018</created>          <gmt_created>2015-12-03 22:00:18</gmt_created>          <changed>1475894869</changed>          <gmt_changed>2016-10-08 02:47:49</gmt_changed>      </item>          <item>          <nid>210141</nid>          <type>image</type>          <title><![CDATA[Robots Reaching Through Clutter - 2]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[kemp_robot2.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/kemp_robot2_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/kemp_robot2_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/kemp_robot2_0.jpg?itok=nsxuCstX]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Robots Reaching Through Clutter - 2]]></image_alt>                    <created>1449180018</created>          <gmt_created>2015-12-03 22:00:18</gmt_created>          <changed>1475894869</changed>          <gmt_changed>2016-10-08 02:47:49</gmt_changed>      </item>          <item>          <nid>210151</nid>          <type>image</type>          <title><![CDATA[Robots Reaching Through Clutter - 3]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[kemp_robot1.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/kemp_robot1_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/kemp_robot1_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/kemp_robot1_0.jpg?itok=bsAC3j8s]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Robots Reaching Through Clutter - 3]]></image_alt>                    <created>1449180018</created>          <gmt_created>2015-12-03 22:00:18</gmt_created>          <changed>1475894869</changed>          <gmt_changed>2016-10-08 02:47:49</gmt_changed>      </item>      </hg_media>  <related>          <link>        <url><![CDATA[http://healthcare-robotics.com/]]></url>        <title><![CDATA[Healthcare Robotics Lab]]></title>      </link>          <link>        <url><![CDATA[http://charliekemp.com/]]></url>        <title><![CDATA[Website of Dr. Charlie Kemp]]></title>      </link>          <link>        <url><![CDATA[http://www.youtube.com/user/HealthcareRobotics]]></url>        <title><![CDATA[Additional Videos on YouTube]]></title>      </link>      </related>  <files>      </files>  <groups>          <group id="1214"><![CDATA[News Room]]></group>      </groups>  <categories>          <category tid="135"><![CDATA[Research]]></category>          <category tid="152"><![CDATA[Robotics]]></category>      </categories>  <news_terms>          <term tid="135"><![CDATA[Research]]></term>          <term tid="152"><![CDATA[Robotics]]></term>      </news_terms>  <keywords>          <keyword tid="2157"><![CDATA[Charlie Kemp]]></keyword>          <keyword tid="594"><![CDATA[college of engineering]]></keyword>          <keyword tid="36141"><![CDATA[Coulter Department of Biomedical Engineering at Georgia Tech and Emory University]]></keyword>          <keyword tid="12319"><![CDATA[Healthcare Robotics Lab]]></keyword>          <keyword tid="65291"><![CDATA[Henry Evans]]></keyword>          <keyword tid="65331"><![CDATA[Meka Robotics]]></keyword>          <keyword tid="65321"><![CDATA[robots reaching in clutter]]></keyword>          <keyword tid="65251"><![CDATA[tactile sensing]]></keyword>      </keywords>  <core_research_areas>          <term tid="39521"><![CDATA[Robotics]]></term>      </core_research_areas>  <news_room_topics>      </news_room_topics>  <files></files>  <related></related>  <userdata><![CDATA[]]></userdata></node><node id="158301">  <title><![CDATA[Home-Based Assessment Tool for Dementia Screening]]></title>  <uid>27560</uid>  <body><![CDATA[<p>With baby boomers approaching the age of 65 and new cases of Alzheimer’s disease expected to increase by 50 percent by the year 2030, Georgia Tech researchers have created a tool that allows adults to screen themselves for early signs of dementia. The home-based computer software is patterned after the paper-and-pencil Clock Drawing Test, one of health care’s most commonly used screening exams for cognitive impairment.</p><p>“Technology allows us to check our weight, blood-sugar levels and blood pressure, but not our own cognitive abilities,” said project leader Ellen Yi-Luen Do. “Our ClockMe System helps older adults identify early signs of impairment, while allowing clinicians to quickly analyze the test results and gain valuable insight into the patient’s thought processes.”</p><p>Georgia Tech’s ClockMe system eliminates the paper trail and computerizes the test into two main components: the ClockReader Application and the ClockAnalyzer Application. Click <a href="http://www.youtube.com/watch?v=MEbWeiKxd0c">here</a> to see a video demo.&nbsp;</p><p>ClockReader is the actual test and is taken with a stylus and computer or tablet. The participant is given a specific time and instructed to draw a clock with numbers and the correct minute and hour hands. Once completed, the sketch is emailed to a clinician, who uses the ClockAnalyzer Application to score the test. The software checks for 13 traits. They include correct placement of numbers and hands without extra markings. People with cognitive impairment frequently draw clocks with missing or extra numbers. Digits are sometimes drawn outside of the clock. The time is often incorrect.</p><p>In addition to scoring automatically and consistently, ClockAnalyzer records the duration of the test and the time between each stroke. The software also replays the drawing in real-time, allowing a clinician to watch the drawing being created to observe any behavior abnormality.&nbsp;</p><p>“The traditional paper-and-pencil test is usually overseen by a technician and later scored by a clinician, who scores the test based only on the finished drawing,” said Do, a professor in Georgia Tech’s Colleges of Computing and Architecture. “By looking at the sketch, the scorer is not able to decipher whether the person struggled to remember certain numbers while drawing the clock. The ClockMe system’s timing software highlights those delays.”</p><p>And, because they’re saved electronically, the drawings can be used to easily compare a person’s cognitive ability progress or regression over time. Do’s research found that traditional tests are often filed in a folder and are rarely used for future comparison.</p><p>The ClockMe system was initially tested at the Emory Alzheimer’s Disease Research Center in Atlanta, where it’s currently being used in addition to the traditional paper-and-pencil test. Despite a lack of computer literacy, all of the elderly patients who used the software during the study said they had no problems with the pen-based, computer technology.</p><p>“For this reason, as well as the ability to send the drawings directly to clinicians for convenient scoring, we envision ClockMe as a viable tool for home-based screening,” said Do. “America’s health care costs are expected to soar as baby boomers become senior citizens. If a screening tool can be used at home, unnecessary trips to clinics can be eliminated and medical expenses can be saved.”</p><p>Do and her colleagues are hoping to commercialize the project in the future. Their research was <a href="http://iospress.metapress.com/content/b0841g7827151q34/?p=f2fca69c7d9b46fb9de44239d5b178df&amp;pi=4">published</a> in September’s Journal of Ambient Intelligence and Smart Environments.</p><p><em>This project is supported by the National Science Foundation (NSF) (Award Number SHB-1117665). The content is solely the responsibility of the principal investigators and does not necessarily represent the official views of the NSF. </em></p><p>&nbsp;</p>]]></body>  <author>Jason Maderer</author>  <status>1</status>  <created>1349115293</created>  <gmt_created>2012-10-01 18:14:53</gmt_created>  <changed>1475896374</changed>  <gmt_changed>2016-10-08 03:12:54</gmt_changed>  <promote>0</promote>  <sticky>0</sticky>  <teaser><![CDATA[Georgia Tech researchers have created a tool that allows adults to screen themselves for early signs of dementia.]]></teaser>  <type>news</type>  <sentence><![CDATA[Georgia Tech researchers have created a tool that allows adults to screen themselves for early signs of dementia.]]></sentence>  <summary><![CDATA[<p>With baby boomers approaching the age of 65 and new cases of Alzheimer’s disease expected to increase by 50 percent by the year 2030, Georgia Tech researchers have created a tool that allows adults to screen themselves for early signs of dementia. The home-based computer software is patterned after the paper-and-pencil Clock Drawing Test, one of health care’s most commonly used screening exams for cognitive impairment.</p>]]></summary>  <dateline>2012-10-02T00:00:00-04:00</dateline>  <iso_dateline>2012-10-02T00:00:00-04:00</iso_dateline>  <gmt_dateline>2012-10-02 00:00:00</gmt_dateline>  <subtitle>    <![CDATA[]]>  </subtitle>  <sidebar><![CDATA[]]></sidebar>  <email><![CDATA[maderer@gatech.edu]]></email>  <location></location>  <contact><![CDATA[<p>Jason Maderer<br />Media Relations<br /><a href="mailto:maderer@gatech.edu">maderer@gatech.edu</a><br />404-385-2966</p>]]></contact>  <boilerplate></boilerplate>  <boilerplate_text><![CDATA[]]></boilerplate_text>  <media>          <item>158271</item>          <item>158281</item>          <item>158291</item>      </media>  <hg_media>          <item>          <nid>158271</nid>          <type>image</type>          <title><![CDATA[ClockMe System 1]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[dscn1345.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/dscn1345_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/dscn1345_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/dscn1345_0.jpg?itok=ZxCt2ioy]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[ClockMe System 1]]></image_alt>                    <created>1449178883</created>          <gmt_created>2015-12-03 21:41:23</gmt_created>          <changed>1475894792</changed>          <gmt_changed>2016-10-08 02:46:32</gmt_changed>      </item>          <item>          <nid>158281</nid>          <type>image</type>          <title><![CDATA[ClockMe System 2]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[dscn1347.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/dscn1347_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/dscn1347_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/dscn1347_0.jpg?itok=ukUk-syR]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[ClockMe System 2]]></image_alt>                    <created>1449178883</created>          <gmt_created>2015-12-03 21:41:23</gmt_created>          <changed>1475894792</changed>          <gmt_changed>2016-10-08 02:46:32</gmt_changed>      </item>          <item>          <nid>158291</nid>          <type>image</type>          <title><![CDATA[Sample Clock Drawing-Impairment]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[dementia-clock.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/dementia-clock_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/dementia-clock_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/dementia-clock_0.jpg?itok=G1hsnkcJ]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Sample Clock Drawing-Impairment]]></image_alt>                    <created>1449178883</created>          <gmt_created>2015-12-03 21:41:23</gmt_created>          <changed>1475894794</changed>          <gmt_changed>2016-10-08 02:46:34</gmt_changed>      </item>      </hg_media>  <related>      </related>  <files>      </files>  <groups>          <group id="1183"><![CDATA[Home]]></group>      </groups>  <categories>          <category tid="137"><![CDATA[Architecture]]></category>          <category tid="135"><![CDATA[Research]]></category>      </categories>  <news_terms>          <term tid="137"><![CDATA[Architecture]]></term>          <term tid="135"><![CDATA[Research]]></term>      </news_terms>  <keywords>          <keyword tid="44881"><![CDATA[Alzheimer&#039;s Disease]]></keyword>          <keyword tid="1945"><![CDATA[Ellen Yi-luen Do]]></keyword>      </keywords>  <core_research_areas>          <term tid="39501"><![CDATA[People and Technology]]></term>      </core_research_areas>  <news_room_topics>      </news_room_topics>  <files></files>  <related></related>  <userdata><![CDATA[]]></userdata></node><node id="156701">  <title><![CDATA[Georgia Tech Creating High-Tech Tools to Study Autism]]></title>  <uid>27560</uid>  <body><![CDATA[<p class="Body1">Researchers in Georgia Tech’s Center for Behavior Imaging have developed two new technological tools that automatically measure relevant behaviors of children, and promise to have significant impact on the understanding of behavioral disorders such as autism.</p><p class="Body1">One of the tools—a system that uses special gaze-tracking glasses and facial-analysis software to identify when a child makes eye contact with the glasses-wearer—was created by combining two existing technologies to develop a novel capability of automatic detection of eye contact. The other is a wearable system that uses accelerometers to monitor and categorize problem behaviors in children with behavioral disorders.</p><p class="Body1">Both technologies already are being deployed in the Center for Behavior Imaging’s (CBI) ongoing work to apply computational methods to screening, measurement and understanding of autism and other behavioral disorders.</p><p class="Body1">Children at risk for autism often display distinct behavioral markers from a very young age. One such marker is a reluctance to make frequent or prolonged eye contact with other people. Discovering an automated way to detect this and other telltale behavioral markers would be a significant step toward scaling autism screening up to much larger populations than are currently reached. This is one goal of the five-year, $10 million “Expeditions” project, funded in fall 2010 by the National Science Foundation under principal investigator and CBI Director Jim Rehg, also a professor in Georgia Tech’s School of Interactive Computing.</p><p class="Body1">The eye-contact tracking system begins with a commercially available pair of glasses that can record the focal point of their wearer’s gaze. Researchers took video of a child captured by a front-facing camera on the glasses, worn by an adult who was interacting with the child. The video was then processed using facial recognition software available from a second manufacturer. Combine the glasses’ hard-wired ability to detect wearer gaze with the facial-recognition software’s ability to detect the child’s gaze direction, and the result is a system able to detect eye contact in a test interaction with a 22-month-old with 80 percent accuracy. The study was conducted in Georgia Tech’s Child Study Lab (CSL), a child-friendly experimental facility richly equipped with cameras, microphones and other sensors.</p><p class="Body1">“Eye gaze has been a tricky thing to measure in laboratory settings, and typically it’s very labor-intensive, involving hours and hours of looking at frames of video to pinpoint moments of eye contact,” Rehg said. “The exciting thing about our method is that it can produce these measures automatically and could be used in the future to measure eye contact outside the laboratory setting. We call these results preliminary because they were obtained from a single subject, but all humans’ eyes work pretty much the same way, so we’re confident the successful results will be replicated with future subjects.” &nbsp;</p><p>The other new system, developed in collaboration with the Marcus Autism Center in Atlanta and Dr. Thomas Ploetz of Newcastle University in the United Kingdom, is a package of sensors, worn via straps on the wrists and ankles, that uses accelerometers to detect movement by the wearer. Algorithms developed by the team analyze the sensor data to automatically detect episodes of problem behavior and classify them as aggressive, self-injurious or disruptive (e.g., throwing objects).</p><p>Researchers first developed the algorithms by putting the sensors on four Marcus clinic staff members who together performed some 1,200 different behavior instances, and the system detected “problem” behaviors with 95 percent accuracy and classified all behaviors with 80 percent accuracy. They then used the sensors with a child diagnosed along the autism spectrum, and the system detected the child’s problem-behavior episodes with 81 percent accuracy and classified them with 70 percent accuracy.</p><p class="Body1">“These results are very promising in leading the way toward more accurate and reliable measurement of problem behavior, which is important in determining whether treatments targeting these behaviors are working,” said CSL Director Agata Rozga, a research scientist in the School of Interactive Computing and co-investigator on the Expeditions award. “Our ultimate goal with this wearable sensing system is to be able to gather data on the child’s behavior beyond the clinic, in settings where the child spends most of their time, such as their home or school. In this way, parents, teachers and others who care for the child can be potentially alerted to times and situations when problem behaviors occur so that they can address them immediately.”</p><p class="Body1">“What these tools show is that computational methods and technologies have great promise and potential impact on the lives of many children and their parents and caregivers,” said Gregory Abowd, Regents’ Professor in the School of Interactive Computing and a prominent researcher in technology and autism. “These technologies we are developing, and others developed and explored elsewhere, aim to bring more effective early-childhood screening to millions of children nationwide, as well as enhance care for those children already diagnosed on the autism spectrum.”</p><p class="Body1">Both technologies were presented in early September at the 14<sup>th</sup> ACM International Conference on Ubiquitous Computing (Ubicomp 2012). Among the other devices under study at CSL are a camera/software system that can track children’s facial expressions and customized speech analysis software to detect vocalization patterns.</p><p class="Body1">For more information on behavioral imaging, visit the Georgia Tech/NSF website on computational behavioral science at <a href="http://www.cbs.gatech.edu/">http://www.cbs.gatech.edu</a>. For information or to volunteer for one of CBI’s ongoing studies, visit the Child Study Lab website at <a href="http://childstudy.hsi.gatech.edu/">http://childstudy.hsi.gatech.edu</a>.</p>]]></body>  <author>Jason Maderer</author>  <status>1</status>  <created>1348568383</created>  <gmt_created>2012-09-25 10:19:43</gmt_created>  <changed>1475896370</changed>  <gmt_changed>2016-10-08 03:12:50</gmt_changed>  <promote>0</promote>  <sticky>0</sticky>  <teaser><![CDATA[]]></teaser>  <type>news</type>  <sentence><![CDATA[]]></sentence>  <summary><![CDATA[<p class="Body1">Researchers in Georgia Tech’s Center for Behavior Imaging have developed two new technological tools that automatically measure relevant behaviors of children, and promise to have significant impact on the understanding of behavioral disorders such as autism.</p>]]></summary>  <dateline>2012-09-25T00:00:00-04:00</dateline>  <iso_dateline>2012-09-25T00:00:00-04:00</iso_dateline>  <gmt_dateline>2012-09-25 00:00:00</gmt_dateline>  <subtitle>    <![CDATA[Innovations will lead to better treatment, assessment for children]]>  </subtitle>  <sidebar><![CDATA[]]></sidebar>  <email><![CDATA[mterraza@cc.gatech.edu]]></email>  <location></location>  <contact><![CDATA[<p class="Body1">Michael Terrazas<br />Assistant Director of Communications<br />College of Computing<a href="mailto:mterraza@cc.gatech.edu"><br />mterraza@cc.gatech.edu</a><br />404-245-0707</p><p>&nbsp;</p>]]></contact>  <boilerplate></boilerplate>  <boilerplate_text><![CDATA[]]></boilerplate_text>  <media>          <item>156711</item>          <item>60510</item>      </media>  <hg_media>          <item>          <nid>156711</nid>          <type>image</type>          <title><![CDATA[Child Study Lab]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[child-study-lab-1_0.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/child-study-lab-1_0_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/child-study-lab-1_0_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/child-study-lab-1_0_0.jpg?itok=_JU-eTXd]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Child Study Lab]]></image_alt>                    <created>1449178872</created>          <gmt_created>2015-12-03 21:41:12</gmt_created>          <changed>1475894792</changed>          <gmt_changed>2016-10-08 02:46:32</gmt_changed>      </item>          <item>          <nid>60510</nid>          <type>image</type>          <title><![CDATA[Gregory Abowd and James Rehg]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[tzo30302.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/tzo30302_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/tzo30302_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/tzo30302_0.jpg?itok=owTbmJlK]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Gregory Abowd and James Rehg]]></image_alt>                    <created>1449176267</created>          <gmt_created>2015-12-03 20:57:47</gmt_created>          <changed>1475894525</changed>          <gmt_changed>2016-10-08 02:42:05</gmt_changed>      </item>      </hg_media>  <related>      </related>  <files>      </files>  <groups>          <group id="1214"><![CDATA[News Room]]></group>      </groups>  <categories>          <category tid="135"><![CDATA[Research]]></category>      </categories>  <news_terms>          <term tid="135"><![CDATA[Research]]></term>      </news_terms>  <keywords>          <keyword tid="6053"><![CDATA[Autism]]></keyword>          <keyword tid="44431"><![CDATA[Child Study Lab]]></keyword>          <keyword tid="397"><![CDATA[children]]></keyword>      </keywords>  <core_research_areas>          <term tid="39501"><![CDATA[People and Technology]]></term>      </core_research_areas>  <news_room_topics>      </news_room_topics>  <files></files>  <related></related>  <userdata><![CDATA[]]></userdata></node><node id="145221">  <title><![CDATA[Gregory Abowd, Marie Thursby Named Regents' Professors]]></title>  <uid>27445</uid>  <body><![CDATA[<p>Georgia Tech has named Gregory Abowd of the College of Computing and Marie Thursby of the Scheller College of Business Regents' Professors.</p><p>The University System of Georgia Board of Regents grants the professorships to outstanding, tenured full professors based on excellence in research and contributions to their professions and institutions.</p><p>"We are very proud of our new Georgia Tech Regents' Professors," said Rafael L. Bras, provost and executive vice president for Academic Affairs. "By definition, they represent the very best among our extraordinary faculty — they are educators, academicians and scholars of the first order."</p><p>Abowd, a distinguished professor in the School of Interactive Computing, arrived at Tech in 1994 and quickly developed a research agenda on the applications of novel technologies in living laboratories such as the classroom and home. Since 2002, much of his research has been devoted to challenges linking information technologies to autism. Abowd established the Atlanta Autism Consortium in August 2008, which helped to unite Atlanta’s many autism-related constituencies.</p><p>He was selected as a Rhodes Scholar in 1986 and, after completing his research studies at the University of Oxford, worked from 1989 to 1992 as a research associate with the Human-Computer Interaction Group in the Department of Computer Science at the University of York in England. From 1992 to 1994, Abowd was a postdoctoral research associate with the Software Engineering Institute and the Computer Science Department at Carnegie Mellon University.</p><p>"I am very grateful for the recognition of this Regents' Professorship, because it recognizes the importance of the work my colleagues, students and I have been doing," Abowd said. "I am also very grateful because this recognition was initiated by my peers, and it is an indication of their respect for what I have accomplished over the years at Georgia Tech."</p><p>Thursby, a research associate of National Bureau of Economics Research since 1987, is currently Hal and John Smith Chair of Entrepreneurship in the Scheller College of Business. Her research on innovation and technology transfer has been used in congressional, National Institute of Health and U.S. National Academies policy discussions.</p><p>One of her major contributions to Tech since arriving in 2002, has been founding the internationally acclaimed Technological Innovation: Generating Economic Returns (TI:GER) program. TI:GER is a collaboration between the Institute and Emory University and has received multiple awards including the Academy of Management Entrepreneurship Pedagogy Award in 2006. The program teams PhD students in science and engineering with MBA and JD students in an experiential curriculum focused on the intersection of technical, legal and business issues and innovation.</p><p>Prior to arriving at Tech, Thursby was the Burton D. Morgan Chair of International<br />Policy and Management at Purdue University and has held faculty appointments at the<br />University of Michigan, Ohio State University, Syracuse University and North Carolina State University.</p><p>"I feel quite honored and grateful," Thursby said. "At Georgia Tech, I get to work with great colleagues and students in an environment that can't help but make one productive."</p>]]></body>  <author>Amelia Pavlik</author>  <status>1</status>  <created>1344529996</created>  <gmt_created>2012-08-09 16:33:16</gmt_created>  <changed>1475896360</changed>  <gmt_changed>2016-10-08 03:12:40</gmt_changed>  <promote>0</promote>  <sticky>0</sticky>  <teaser><![CDATA[Georgia Tech has named Gregory Abowd of the College of Computing and Marie Thursby of the Scheller College of Business Regents' Professors.]]></teaser>  <type>news</type>  <sentence><![CDATA[Georgia Tech has named Gregory Abowd of the College of Computing and Marie Thursby of the Scheller College of Business Regents' Professors.]]></sentence>  <summary><![CDATA[<p class="Body1">Georgia Tech has named Gregory Abowd of the College of Computing and Marie Thursby of the Scheller College of Business Regents' Professors.</p>]]></summary>  <dateline>2012-08-09T00:00:00-04:00</dateline>  <iso_dateline>2012-08-09T00:00:00-04:00</iso_dateline>  <gmt_dateline>2012-08-09 00:00:00</gmt_dateline>  <subtitle>    <![CDATA[]]>  </subtitle>  <sidebar><![CDATA[]]></sidebar>  <email><![CDATA[]]></email>  <location></location>  <contact><![CDATA[<p><a href="mailto:victor.rogers@comm.gatech.edu">Victor Rogers</a><br />Office of the Provost</p>]]></contact>  <boilerplate></boilerplate>  <boilerplate_text><![CDATA[]]></boilerplate_text>  <media>          <item>145201</item>          <item>145191</item>      </media>  <hg_media>          <item>          <nid>145201</nid>          <type>image</type>          <title><![CDATA[Gregory Abowd]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[abowd.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/abowd_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/abowd_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/abowd_0.jpg?itok=OapN6WNb]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Gregory Abowd]]></image_alt>                    <created>1449178739</created>          <gmt_created>2015-12-03 21:38:59</gmt_created>          <changed>1475894777</changed>          <gmt_changed>2016-10-08 02:46:17</gmt_changed>      </item>          <item>          <nid>145191</nid>          <type>image</type>          <title><![CDATA[Marie Thursby]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[thursby.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/thursby_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/thursby_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/thursby_0.jpg?itok=lol8jsiE]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[Marie Thursby]]></image_alt>                    <created>1449178739</created>          <gmt_created>2015-12-03 21:38:59</gmt_created>          <changed>1475894777</changed>          <gmt_changed>2016-10-08 02:46:17</gmt_changed>      </item>      </hg_media>  <related>      </related>  <files>      </files>  <groups>          <group id="1259"><![CDATA[Whistle]]></group>      </groups>  <categories>          <category tid="129"><![CDATA[Institute and Campus]]></category>      </categories>  <news_terms>          <term tid="129"><![CDATA[Institute and Campus]]></term>      </news_terms>  <keywords>          <keyword tid="654"><![CDATA[College of Computing]]></keyword>          <keyword tid="11002"><![CDATA[Gregory Abowd]]></keyword>          <keyword tid="13925"><![CDATA[Marie Thursby]]></keyword>          <keyword tid="40061"><![CDATA[Regents&#039; Professors]]></keyword>          <keyword tid="167089"><![CDATA[Scheller College of Business]]></keyword>          <keyword tid="166848"><![CDATA[School of Interactive Computing]]></keyword>      </keywords>  <core_research_areas>      </core_research_areas>  <news_room_topics>      </news_room_topics>  <files></files>  <related></related>  <userdata><![CDATA[]]></userdata></node><node id="70408">  <title><![CDATA[Georgia Tech Researchers Receive Three NSF Emerging Frontiers Awards]]></title>  <uid>27174</uid>  <body><![CDATA[<p>The National Science Foundation (NSF) has awarded $6 million to fund three projects involving researchers from the Georgia Institute of Technology. Each four-year, $2 million grant was awarded through the NSF's Division of Emerging Frontiers in Research and Innovation (EFRI). </p><p>"The EFRI research teams will probe some profound aspects of the interface of biology and engineering," said Sohi Rastegar, director of EFRI. "If they are successful, the principles and theories uncovered in their investigations could unlock many technological opportunities."</p><p>This year, 14 transformative, fundamental research projects were awarded EFRI grants in two emerging areas: technologies that build on understanding of biological signaling, and machines that can interact and cooperate with humans. </p><p>The three Georgia Tech projects include:</p><ul><li>Developing a "therapeutic robot" to help rehabilitate and improve motor skills in people with mobility problems;</li><li>Creating wearable sensors that allow blind people to "see" with their hands, bodies or faces;</li><li>Generating and rigorously testing quantitative models that describe spatial and temporal regulation of cell differentiation in tissues.</li></ul><p>The therapeutic robot could enhance, assist and improve motor skills in humans with varying motor capabilities and deficits. The goal of the project is to program a humanoid rehabilitation robot to perform a "partnered box step," which is a defined pattern of weight shifts and directional changes, solely based on interpreting movement cues from subtle changes in forces between the hands and arms of the robot and the person. </p><p>To do this, researchers at Georgia Tech and Emory University will study how humans use their muscles to walk, balance and generate force signals with the hands for guidance when moving in cooperation with another person. They will also study "rehabilitative partnered dance," which has been specifically adapted to help improve gait and balance in individuals with motor impairments. </p><p>"Our vision is to develop robots that will interact with humans as both assistants and movement therapists," explained principal investigator Lena Ting, an associate professor in the Wallace H. Coulter Department of Biomedical Engineering at Georgia Tech and Emory University. "We expect our project to have a long-term impact on quality of life of individuals with movement difficulties, such as those caused by Parkinson's disease, stroke and injury by improving fitness, motor skills and social engagement."</p><p>Working with Ting on the project are Emory University School of Medicine (geriatrics) assistant professor Madeleine Hackney, Coulter Department of Biomedical Engineering assistant professor Charlie Kemp and Georgia Tech School of Interactive Computing assistant professor Karen Liu.</p><p>For the second project, researchers at Georgia Tech and The City College of New York will investigate devices for "alternative perception" and the principles underlying the human-machine interaction. Alternative perception combines electronics and the other senses to emulate vision. In addition to aiding the visually impaired, the findings are expected to have other applications, such as the development of intelligent robots. </p><p>The researchers plan to untangle how humans learn to coordinate input from their senses -- e.g. vision, touch -- with movements, like reaching for a glass or moving through a crowded room. They will then map out how machines, such as robots and computers, learn similar tasks, to model devices that can assist humans. </p><p>The team envisions a multifunctional array of sensors on the body and has already developed prototypes for some of the devices. The full complement of wearable sensors would help a sightless person navigate by conveying information about his or her surroundings. </p><p>The researchers hope their findings on perception, and the prototypes they develop, will spawn a raft of wearable electronic devices to help blind people "see" their environment at a distance through touch, hearing and other senses. The technology would also benefit sighted individuals who must navigate in poor visibility, such as firefighters and pilots.</p><p>Principal investigator Zhigang Zhu, professor of computer science and computer engineering in City College's Grove School of Engineering, will collaborate with City College professor of psychology and director of the Program in Cognitive Neuroscience Tony Ro, City College professor of electrical engineering Ying Li Tian, Georgia Tech Woodruff School of Mechanical Engineering professor Kok-Meng Lee, and Georgia Tech School of Applied Physiology associate professor Boris Prilutsky.</p><p>The third project will address a fundamental question of developmental biology: what controls the spatial and temporal patterns of cell differentiation? Answering this question will lead to a better understanding of the basic principles of embryogenesis, explain origins of developmental disorders, and provide guidelines for tissue engineering and regenerative medicine. </p><p>The research will be conducted by principal investigator and Princeton University Department of Chemical and Biological Engineering associate professor Stanislav Shvartsman, Georgia Tech School of Chemical and Biomolecular Engineering associate professor Hang Lu, New York University Department of Biology professor Christine Rushlow, and University of Illinois at Urbana Champaign Department of Computer Science associate professor Saurabh Sinha.</p><p>Scientists know that among an embryo's first major developments is the establishment of its dorsoventral axis, which runs from its back to its belly. The researchers plan to study how this axis development unfolds -- specifically the presence and location of proteins during the process, which give rise to muscle, nerve and skin tissues. </p><p>To enable large-scale quantitative analyses of protein positional information along the dorsoventral axis, Lu and Shvartsman will further develop a microfluidic device they previously designed to reliably and robustly orient several hundred embryos in just a few minutes.</p><p>"By understanding this system at a deeper, quantitative level, we will elucidate general principles underlying the operation of genetic and multicellular networks that drive development," said Lu.</p><p><strong>Research News &amp; Publications Office<br />Georgia Institute of Technology<br />75 Fifth Street, N.W., Suite 314<br />Atlanta, Georgia 30308 USA</strong></p><p><strong>Media Relations Contacts:</strong> Abby Robinson (abby@innovate.gatech.edu; 404-385-3364) or John Toon (jtoon@gatech.edu; 404-894-6986)</p>]]></body>  <author>Mike Terrazas</author>  <status>1</status>  <created>1317308368</created>  <gmt_created>2011-09-29 14:59:28</gmt_created>  <changed>1475896214</changed>  <gmt_changed>2016-10-08 03:10:14</gmt_changed>  <promote>0</promote>  <sticky>0</sticky>  <teaser><![CDATA[]]></teaser>  <type>news</type>  <sentence><![CDATA[]]></sentence>  <summary><![CDATA[<p>The National Science Foundation has awarded $6 million through its Division of Emerging Frontiers in Research and Innovation to fund three projects involving researchers from Georgia Tech, including Karen Liu and Charlie Kemp (<em>Interactive Computing</em>). <em>Source: GT Research News</em></p>]]></summary>  <dateline>2011-09-29T00:00:00-04:00</dateline>  <iso_dateline>2011-09-29T00:00:00-04:00</iso_dateline>  <gmt_dateline>2011-09-29 00:00:00</gmt_dateline>  <subtitle>    <![CDATA[]]>  </subtitle>  <sidebar><![CDATA[]]></sidebar>  <email><![CDATA[]]></email>  <location></location>  <contact><![CDATA[]]></contact>  <boilerplate></boilerplate>  <boilerplate_text><![CDATA[]]></boilerplate_text>  <media>          <item>70367</item>      </media>  <hg_media>          <item>          <nid>70367</nid>          <type>image</type>          <title><![CDATA[(L-R) Lena Ting, Karen Liu, Charlie Kemp and Madeleine Hackney]]></title>          <body><![CDATA[]]></body>                      <image_name><![CDATA[tinggroup195.jpg]]></image_name>            <image_path><![CDATA[/sites/default/files/images/tinggroup195_0.jpg]]></image_path>            <image_full_path><![CDATA[http://hg.gatech.edu//sites/default/files/images/tinggroup195_0.jpg]]></image_full_path>            <image_740><![CDATA[http://hg.gatech.edu/sites/default/files/styles/740xx_scale/public/sites/default/files/images/tinggroup195_0.jpg?itok=XB1ZuUAD]]></image_740>            <image_mime>image/jpeg</image_mime>            <image_alt><![CDATA[(L-R) Lena Ting, Karen Liu, Charlie Kemp and Madeleine Hackney]]></image_alt>                    <created>1449177304</created>          <gmt_created>2015-12-03 21:15:04</gmt_created>          <changed>1475894618</changed>          <gmt_changed>2016-10-08 02:43:38</gmt_changed>      </item>      </hg_media>  <related>      </related>  <files>      </files>  <groups>          <group id="47223"><![CDATA[College of Computing]]></group>      </groups>  <categories>      </categories>  <news_terms>      </news_terms>  <keywords>          <keyword tid="1102"><![CDATA[blind]]></keyword>          <keyword tid="14478"><![CDATA[Boris Prilutsky]]></keyword>          <keyword tid="14480"><![CDATA[cell differentiation]]></keyword>          <keyword tid="2157"><![CDATA[Charlie Kemp]]></keyword>          <keyword tid="654"><![CDATA[College of Computing]]></keyword>          <keyword tid="594"><![CDATA[college of engineering]]></keyword>          <keyword tid="11533"><![CDATA[Department of Biomedical Engineering]]></keyword>          <keyword tid="898"><![CDATA[Hang Lu]]></keyword>          <keyword tid="2296"><![CDATA[Karen Liu]]></keyword>          <keyword tid="14477"><![CDATA[Kok-Meng Lee]]></keyword>          <keyword tid="2266"><![CDATA[Lena Ting]]></keyword>          <keyword tid="7341"><![CDATA[microfluidic]]></keyword>          <keyword tid="1482"><![CDATA[mobility]]></keyword>          <keyword tid="1356"><![CDATA[robot]]></keyword>          <keyword tid="167863"><![CDATA[School of Applied Physiology]]></keyword>          <keyword tid="167445"><![CDATA[School of Chemical and Biomolecular Engineering]]></keyword>          <keyword tid="166848"><![CDATA[School of Interactive Computing]]></keyword>          <keyword tid="167377"><![CDATA[School of Mechanical Engineering]]></keyword>          <keyword tid="167318"><![CDATA[sensor]]></keyword>          <keyword tid="14479"><![CDATA[therapeutic robot]]></keyword>      </keywords>  <core_research_areas>      </core_research_areas>  <news_room_topics>      </news_room_topics>  <files></files>  <related></related>  <userdata><![CDATA[]]></userdata></node></nodes>