Alle Angebote für dieses Produkt
Über dieses Produkt
- KurzbeschreibungLeonardo da Vinci once said,
The eye encompasses the beauty of the whole world.
Making this visual beauty of the world more accessible to visually impaired people has inspired researchers in Computer Vision for a long time. Perhaps the most ambitious software solution for the vision problem would be an algorithm that produces a semantic description of the image content which is then output of a speech synthesis device in natural language. This automated image analysis system would mimic a partner with normal vision who describes the image to the user. However, despite the fact that automated image understanding will remain a challenge to researchers for many years, it would continue to deprive the visually impaired of a direct perceptual experience, an active exploration, and an impression of where things are in the image and what visual appearance they have. The approaches, described in this work, therefore, are to augment the sensory capabilities of visually impaired persons by translating image content into sounds. The task of analyzing and understanding images is still up to the user, which is why we call our approach &8220;auditory image understanding&8221;. Very much like a blind person who explores a Braille text or a bas-relief image haptically with the tip of her finger, our users touch the image via touch screen and experience the local properties of the image as auditory feedback.
The approaches presented combine low-level information, such as color, edges, and roughness, with mid- and high-level information obtained from Machine Learning algorithms. This includes object recognition and the classification of regions into the categories &8220;man-made&8221; versus &8220;natural&8221; based on a novel type of discriminative graphical model. We argue that this multi-level approach gives users direct access to the identity and location of objects and structures in the image, yet it still exploits the potential of recent developments in Computer Vision and Machine Learning. The sound translation is inspired by the way humans perceive colors and structures visually.
Due to the simplicity and directness of the sensory mapping from visual to auditory, we harness the human ability to learn, so we consider the brain of the user as a fundamental part of the system. Visually impaired persons can use the system to analyze images that they find on the internet, but also for personal photos that their friends or loved ones want to share with them. It is this application scenario that makes the direct perceptual access most valuable
- AutorMichael Banf
- Ausgabe1. Auflage
- Seiten300 Seiten
- Gewicht450 g
Meistverkauft in Sprache & Literatur
Hier sparen: Sprache & Literatur
- EUR 15,00Preistendenz: EUR 15,30
- EUR 22,00Preistendenz: EUR 22,88
- EUR 14,50Preistendenz: EUR 14,79
- EUR 19,99Preistendenz: EUR 20,38
- EUR 19,95Preistendenz: EUR 20,64
- EUR 9,95Preistendenz: EUR 10,28
- EUR 40,00Preistendenz: EUR 41,59
Dieser Artikel gehört nicht auf diese Seite.
Vielen Dank. Wir kümmern uns darum.