© Universität Bielefeld

Faculty of Technology: News

Faculty of Technology: News - Tag [citec]

How the Brain Integrates Sensory Input

Veröffentlicht am 3. May 2019
Bielefeld researchers publish study on the flexibility of sensory perception 

Hearing, sight, touch – our brain captures a wide range of distinct sensory stimuli and links them together. The brain has a kind of built-in filter function for this: sensory impressions are only integrated if it is necessary and useful for the task at hand. Researchers from Bielefeld University, Oxford University (Great Britain), and Aix-Marseille University (France) investigated this phenomenon of flexibility in perception, and have now published a study on their findings that appears in the scientific journal “Neuron” (29 April 2019). In their publication, the researchers reveal where sensory stimuli are integrated in the brain, and in which area of the brain this flexibility can be located. From Bielefeld University, Professor Dr. Christoph Kayser and Dr. Hame Park from the Cluster of Excellence CITEC were involved in the study.


Prof. Dr. Christoph Kayser. Photo: CITEC/ Universität Bielefeld
Prof. Dr. Christoph Kayser. Photo: CITEC/ Universität Bielefeld
[Weiterlesen]
Gesendet von ALauterbach in Forschung
Tags: citec

Following nature's example: animals are prototypes for robotics

Veröffentlicht am 11. July 2018
The natural way in which living beings orient themselves and move in their environment, avoid obstacles, and find their way home again without their brain consuming much energy is a model for scientists who also want to equip robots with such complex abilities. Because in this field there is a wide gulf between the brain and electronics. The exchange of expertise between behavioural neurologists and robotics scientists is the focus of a symposium at the FENS Forum 2018 in Berlin (7.-11.7. 2018).

Bumblebees have a small brain but they cover considerable distance in the search for food. Depending on the species, their radius is up to three kilometres. The flight route is teeming with enemies and obstacles. Changing wind speeds and wind directions add to the hazards. The insects have to steer their flight through a changeable environment, navigate extensively and learn how to find a good source of food and get home to their nests.

"When bumblebees leave their nest for the first time, they take flights to learn their surroundings so that they can find their way back," says Dr. Olivier Bertrand from the Department of Neurobiology at Bielefeld University, Germany. "These flights have a loop-like pattern, whereby the pattern varies from animal to animal, as our studies show. We assume that the bumblebees store snapshots of their environment in their brain, the usefulness of which is checked on subsequent flights."

When flying within a complex cluttered environment, bees constantly need to evaluate the environmental features and have to make decisions that influence the flight course. Dr. Shridar Ravi from the RMTI University in Melbourne, Australia, used bumblebees to seek insights into the mechanisms used for gap identification when the bees are confronted with an obstacle in their flight path and have to assess gap properties. Bees spend significant time in the near vicinity of the gap while performing rapid lateral maneuvers and looking at the gap, as if they would scan the gap to collect important information. In doing so the bee could detect the edges of the gap by utilising the difference between the relative motion of the gap edges and the foreground or background: a closer object moves relative faster than objects in the background.

As long as the capabilities of robots are limited, linking the abilities of animals with those of robots could be helpful. The team led by Prof. Dr. Noriyasu Ando from the Research Center for Advanced Science and Technology in Tokyo has taken this path: they have developed an insect-driven mobile robot. "A male silkmoth sits in a cockpit and his walking controls the robot
 
and directs it to a female moth as soon as he notices her sexual pheromone and reacts to it", is how Professor Ando describes the principle.
"From a technical point of view, this hybrid robot’s performance matches our goal: the future insect mimetic robot will have the model of the insect brain." The hybrid robot also provides scientists with insights into the behaviour of insects. By changing the sensory input and/or the motor output of the robot, the team was able to uncover the sensory-motor control of the reactions of silkmoths to odours. "The hybrid robot enables us to compare an insect brain with an electronic model," said Professor Ando. "Now the robot is controlled directly by a real silkmoth. If the insect is replaced by a robot model of this insect, we can directly compare the performance of the insect brain with that of the model brain on this robot platform. It's still a conceptual idea, but we're working on it.”

Die Forschung von Professorin Dr. Elisabetta Chicca ebnet den Weg für die autonome Navigation.
Prof. Dr. Elisabetta Chicca
Neuromorphic systems are similar to the neuronal networks of the brain. Their hardware is highly specialized and highly interconnected. A team led by Prof. Dr. Elisabetta Chicca of Bielefeld University has developed a neuromorphic model that will enable autonomous mobile systems to navigate better and avoid obstacles in complex environments. The "laser eyes" (laser rangefinder) of autonomous cars detect obstacles, but are very expensive despite many years of development. They consume a lot of energy and - as current incidents have shown - misinterpretations occur in certain situations. The system of the Bielefeld scientists could bring progress in this area.
"We have developed a new electronic motion detector, the “Spiking Elementary Motion Detector”, which can detect the relative motion of objects”, says Professor Chicca. Every car or train driver knows what a "relative movement" is: the church tower in the distance glides slowly past, while the tree at the roadside rushes very quickly past. Insects use such information during navigation in the terrain to avoid collisions.
The new motion detector, sEMD for short, is a technical nerve cell with an artificial synapse. It can pick up signals and produce signals when two pulses arrive within a certain time - hence the name suffix "spiking". A chip can carry thousands of these detectors, depending on the experiment.

The detectors receive their input from innovative neuromorphic cameras, developed by a company in switzerland. In contrast to normal cameras, the pixels of the sensors in these cameras only produce a signal independently if something changes in their "field of vision". These signals are picked up by the motion detector's receptive fields. Each detector has two receptive fields, each receiving signals from nine pixels. If more than half of the pixels of a receptive field are activated, the receptive field produces a signal that is further processed by the detector. The detector can calculate the relative speed at which an object moves in front of the camera based on the time intervals between the signals of two adjacent receptive fields. "Our experiments show that it is possible to generate information for the navigation of robots that avoid collisions," explains Professor Chicca. "Our results pave the way for the construction of low-power compact systems for autonomous navigation. In addition, the sEMD is a universally applicable element for calculating time differences and can therefore also be used for processing other sensory stimuli, for example for locating the source of a sound.


Contact
Dr. Olivier Bertrand
Universität Bielefeld, Neurobiologie
https://ekvv.uni-bielefeld.de/pers_publ/publ/PersonDetail.jsp?personId=34427052

Prof. Dr. Elisabetta Chicca
Universität Bielefeld, AG Neuromorphic Behaving Systems
https://ekvv.uni-bielefeld.de/pers_publ/publ/PersonDetail.jsp?personId=26461080
Gesendet von SKonermann in Forschung

Team from Cluster of Excellence CITEC prevails in RoboCup finale

Veröffentlicht am 25. June 2018
Team from Cluster of Excellence CITEC prevails in RoboCup finale

A team of students and researchers from the Cluster of Excellence Cognitive Interaction Technology (CITEC) at Bielefeld University won the RoboCup World Championship in Montreal, Canada. RoboCup is the leading, and largest, competition for intelligent robots in the world. The “Team of Bielefeld” (ToBi) showed its skills with Pepper the robot in the household service league. More than 400 teams from around the world competed in the various leagues of the competition from 18-22 June 2018. The researchers are now back in Bielefeld.

Das CITEC-Team erreichte mit Pepper bei der RoboCup-Weltmeisterschaft den ersten Platz in der Haushaltsliga. Foto: Universität Bielefeld/CITEC
The CITEC team earned first place with Pepper in the household service league of the RoboCup World Championship. Photo: Bielefeld University/CITEC

Dr. Sven Wachsmuth, who heads the CITEC Central Labs, and his research associate Florian Lier led the team, together with Master’s student Johannes Kummert. “It’s fantastic to see how the students have progressed from the first preparations to the competition,” says Wachsmuth. “They learned to deal with complex systems like robots, and to work independently with them. That we were then able to take first place is, of course, a great success.”

Lier adds: “The team prepared itself very well, also for dealing with uncertainties. The infrastructure there is different from that in the lab. The students put a lot of work into making the software as stable as possible, and they succeeded in this.”

In the household service league RoboCup@Home, their robot had to master various assistive tasks as precisely as possible, including working as a waiter, bringing groceries into the home, loading a dishwasher, giving visitors an introductory tour of RoboCup, and finding its way in unfamiliar surroundings. The CITEC team competed in the Social Standard Platform League (SSPL), a subleague of the household service league. In the SSPL, teams only compete with Pepper, a robot produced by the company Softbank. Second place went to the team from Australia, where the next RoboCup competition will be held, and the team from Chile took third place.

Student Janneke Simmering from the CITEC team took part in the robot world championship for the first time. “The exciting question was: will the robot do what it’s supposed to do? We spent four weeks programming the software and tried to prepare for as many factors and eventualities as possible. The work paid off, and that’s a great feeling. We’re celebrating now.”

Members of this year’s team included: Robert Feldhans, Felix Friese, Kai Konen, David Leins, Jan Patrick Nülle, Sarah Schröder, Janneke Simmering, Philipp von Neumann-Cosel, Johannes Kummert, Florian Lier and Sven Wachsmuth. The preparations for RoboCup are incorporated into a university seminar – each year, new students from the course work together in the team. The Cluster of Excellence Cognitive Interaction Technology (CITEC) has participated in RoboCup since 2009. In 2016, the team earned the title of world champion for the first time, and the team has also taken third place a total of three times: 2012, in Mexico; 2015, in China; and 2017, in Japan.

The Cluster of Excellence Cognitive Interaction Technology (CITEC) at Bielefeld University is one of 43 clusters of excellence in Germany, and the only cluster with a focus in robotics. CITEC is working to make technical systems intuitive and easy to operate. CITEC’s interdisciplinary approach combines cognitive research with technology. Since 2007, CITEC has been funded as part of the Excellence Initiative of the German federal and state governments. Approximately 250 researchers work at the Cluster.
Gesendet von SKonermann in Forschung

Tag Hinweis

Auf dieser Seite werden nur die mit dem Tag [citec] versehenen Blogeinträge gezeigt.

Wenn Sie alle Blogeinträge sehen möchten klicken Sie auf: Startseite

Kalender

« April 2024
SunMonTueWedThuFriSat
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
    
       
Heute