Interview with Hanna Lukashevich, Fraunhofer IDMT
“Multimodal design must follow a less-is-more philosophy”
After completing her degree in physics at the Belarusian State University, Lukashevich joined Fraunhofer IDMT in Ilmenau, Thuringia, Germany.
Fraunhofer IDMT
As audio-driven interfaces expand in next-gen cabins, lighting must stay precise, safe and tightly synchronised. Hanna Lukashevich from Fraunhofer IDMT explains how audio-reactive lighting evolves into robust, multimodal interior systems.
Hanna Lukashevich has been with Fraunhofer IDMT for more
than eighteen years and has led its Semantic Music Technologies department for
over eleven of them. Her work spans AI-based vocal detection, musical
meta-feature extraction and the transfer of semantic music-analysis methods to
industrial audio — demonstrating how techniques
developed for music can be applied effectively to automotive use cases.
At the intersection of sound,
lighting and intelligent cabin sensing, she offers key insights into the
next generation of multimodal in-vehicle experiences. Building on this
expertise, we spoke with the scientist about the technological demands shaping
future audio-reactive interior lighting.
ADT: We are in the midst of a dynamic and
disruptive decade for the automotive industry. From your perspective, what are
the biggest challenges the interior lighting sector will face over the next
five years?
Lukashevich: A key trend is clear: simplification.
Less is more will continue to shape interior lighting, driven by strong cost
pressure and the need to manage increasing technical complexity. At the same
time, safety considerations will gain importance. The debate around distraction
by light will grow, meaning that intensity, motion and dwell times will require
conservative and well-defined limits. From an audio-technology perspective, an
additional set of challenges becomes crucial when lighting reacts to sound. The
most difficult aspects over the next five years lie exactly at the interface
between audio and light. First, achieving low latency and tight synchronization
is essential. For beat- or voice-driven cues to feel natural, end-to-end
detection to LED response must reliably stay within a few tens of milliseconds,
even as the number of addressable LEDs increases. Second, designing balanced
visualizations across audio genres is becoming more demanding. Lighting that
simply follows bass works for certain music styles but fails for speech,
classical or ambient content. Future systems must react to richer audio
features such as tempo, loudness, timbre and mood, and adapt color palettes and
motion accordingly. Third, finding the right trade-off between responsiveness
and stability is critical. Over-sensitive mappings cause flicker during rapid
audio changes, while heavily smoothed ones lose rhythm and emotional impact.
Solutions lie in adaptive smoothing, tempo-grid alignment and guardrails for
color and motion transitions.
What additional considerations come into play?
Beyond audio-reactive lighting, interior lighting must be
treated as a core design element early in the development process. Scaling up
introduces its own technical demands: per-LED calibration and thermal
compensation to maintain uniform whites and hues; low-noise electronics to
prevent dense LED buses from interfering with microphones or audio DSP;
power-aware animations that respect EV energy budgets; privacy-by-design for
any in-cabin audio analytics; and forward compatibility with emerging domain or
Ethernet lighting architectures. With today’s protocol fragmentation, momentum
toward standardization can also be expected, both in communication protocols
and shared application-level semantics, to ensure interoperable and safe
behavior.
Your research covers audio signal processing, machine
learning and semantic technologies. How can insights from this work inspire
innovation in automotive interiors, particularly in creating multimodal user
experiences?
Audio signal processing, machine learning and semantic
technologies enable a shift from simple reactive color effects to truly
authored multimodal in-cabin experiences. Real-time music analysis, capturing
beat, tempo, loudness, timbre, mood and speech-music transitions, allows
lighting to respond in a way that enhances the atmosphere without distracting
the driver. Object-based audio adds spatial context; by linking individual
sound objects to specific lighting zones, vocals, drums or effects can gain subtle
visual counterparts that increase immersion and help passengers perceive the
scene more clearly. A semantic layer, combining information such as time, route
context or user preferences, turns these features into coherent lighting scenes
with consistent choices for color, motion and intensity that remain robust
across genres. To push quality further while keeping latency low, a hybrid
pipeline is promising. Pairing real-time control with AI-driven offline
pre-authoring enables the generation of visualizations in advance, which can be
stored as lighting descriptors and retrieved via audio fingerprinting when a
known track plays. This delivers reliable and high-quality behavior across
different music styles while avoiding typical issues such as oversensitivity or
sluggish smoothing. Throughout all of this, multimodal design must follow a
less-is-more philosophy: control complexity and cost, and always prioritize
safety by avoiding rapid contrast changes or patterns that could distract the
driver.
Fraunhofer IDMT is working on anomaly detection and
audio-visual content verification. How could such technologies contribute to
smarter interior systems, for example by linking sound, sensing and adaptive
lighting to improve safety and comfort?
Fraunhofer IDMT’s work on anomaly detection and audio-visual
content verification can make interior systems smarter by linking sound,
sensing and adaptive lighting to enhance both safety and comfort. Environmental
sound analysis allows the system to detect and localize hazards such as sirens,
horns or approaching two-wheelers and to respond within tens of milliseconds.
For example, the vehicle can briefly attenuate entertainment audio and guide
attention with subtle, zone-specific light cues on the side of the detected
event, ensuring that the driver notices the hazard without being distracted.
The same tools can monitor mechanical noises inside the vehicle. Wear-related
sounds such as rattling, squeaking, banging, bearing whine or pump cavitation
can be transformed into calm, localized maintenance prompts instead of
intrusive alerts, helping prevent failures and reducing complaints related to
noise, vibration and harshness. Audio-visual content verification further
enhances system trust. It can detect spoofed or synthetic voice commands; if
confidence is low, the system can block the action, request a simple
confirmation and indicate this state with a distinct verification needed light
pattern. All processing should prioritize on-device execution using minimal,
short-lived features to protect privacy. Lighting responses must remain clear,
consistent and non-overloading to ensure safety and comfort at all times.