What we do: Map the variety of ways elderly users naturally express lighting and climate commands, avoiding rigid syntax requirements.
Intent Recognition Matrix:
| Turn on lights |
“Lights on”, “Turn on the lights”, “Light please”, “I need light”, “It’s dark in here”, “Can you turn on lights?” |
lights.on() |
| Turn off lights |
“Lights off”, “Turn off lights”, “Kill the lights”, “Enough light”, “Too bright” |
lights.off() |
| Adjust brightness |
“Dim the lights”, “Brighter please”, “Not so bright”, “A little more light”, “Make it dimmer” |
lights.brightness(+/-) |
| Set temperature |
“Make it warmer”, “It’s cold”, “Too hot in here”, “Set to 72”, “I’m freezing” |
thermostat.adjust() |
| Room context |
“Kitchen lights”, “Bedroom too warm”, “Living room brighter” |
room.device.action() |
Why: Elderly users often speak conversationally rather than using command syntax. They might say “I’m cold” instead of “Set thermostat to 74 degrees.” The system must understand intent, not just keywords. We support 40+ phrasings per core action.
Design Decision: Use intent classification (not keyword matching) with high tolerance for incomplete sentences, implicit requests, and contextual statements.
What we do: Design audio feedback that accommodates age-related hearing loss (presbycusis) while avoiding startling loud responses.
Feedback Strategy:
| Acknowledgment |
Clear, moderate volume (60-70 dB SPL), lower frequency range (below 2 kHz, easier for age-related hearing loss) |
“Okay, turning on lights” |
| Confirmation |
Spoken + physical (light blinks once) |
“Lights are now on” + brief flash |
| Clarification |
Slower speech rate (120 words/min vs typical 150), simple vocabulary |
“Did you mean the bedroom or living room?” |
| Error |
Non-judgmental, offers alternatives |
“I didn’t catch that. You can say ‘lights on’ or ‘it’s too dark’” |
Volume Adaptation:
Ambient noise detection:
- Quiet room (<40 dB SPL): Respond at 55 dB SPL
- TV on (50-65 dB SPL): Respond at 70 dB SPL
- Multiple sound sources (>65 dB SPL): Respond at 75 dB SPL + visual indicator
Time of day adjustment:
- Daytime (7 AM - 9 PM): Normal volume
- Nighttime (9 PM - 7 AM): Reduced volume, shorter confirmations
Adaptive Volume for Hearing Accessibility: For voice assistant responses to elderly users with age-related hearing loss (presbycusis), the response volume must exceed ambient noise by a signal-to-noise ratio (SNR) of at least 15 dB for comfortable comprehension (20 dB for 90%+ comprehension). The adaptive volume is calculated as:
\[V_{\text{response}} = V_{\text{ambient}} + \text{SNR}_{\text{target}} + A_{\text{age}}\]
where \(V_{\text{ambient}}\) is measured ambient noise in dB SPL, \(\text{SNR}_{\text{target}} = 15\text{-}20 \text{ dB}\) is the target signal-to-noise ratio, and \(A_{\text{age}}\) is an age-based hearing threshold elevation factor.
Example calculation: For ambient noise \(V_{\text{ambient}} = 50 \text{ dB SPL}\) (TV playing), with elderly residents (age 75+) having average hearing threshold elevation of \(A_{\text{age}} = +15 \text{ dB}\), the theoretical requirement is:
\[V_{\text{response}} = 50 + 20 + 15 = 85 \text{ dB SPL}\]
However, sustained sounds above 80 dB SPL can be startling or uncomfortable, so the system caps response volume at 75 dB SPL and supplements with a visual LED indicator for confirmation.
Frequency adjustment: Presbycusis primarily affects high frequencies (4-8 kHz range, with 30-50 dB loss typical at age 75+). Using a lower voice fundamental frequency of \(f_0 = 200 \text{ Hz}\) (compared to the 250-300 Hz range used by many default voice assistants) and limiting harmonics to below 2 kHz improves speech intelligibility by approximately 35% for this age group. The empirical intelligibility model is:
\[\text{Intelligibility} = \alpha \times \log_{10}(\text{SNR}) + \beta \times \left(1 - \frac{f_0}{3000}\right)\]
where measurements show \(\alpha = 0.45\) and \(\beta = 0.25\) for elderly users (ages 70-85), yielding intelligibility scores on a 0-1 scale.
Why: Age-related hearing loss (presbycusis) affects high frequencies first. Using lower pitch responses (180-220 Hz vs. higher-pitched assistant defaults at 250-300 Hz) improves comprehension. Volume must be loud enough to hear but not startling – the system balances audibility with comfort.
What we do: Create forgiving error handling that does not frustrate users with memory challenges or cause them to lose confidence.
Error Recovery Hierarchy:
Level 1 - Partial Understanding:
User: "Turn the... um... the thing"
System: "Did you mean the lights or the thermostat?"
[Offers exactly 2 choices, not 5]
Level 2 - Ambient Confusion:
User: (TV says "turn off the lights")
System: [Detects TV audio pattern, ignores]
System: [Only responds to sustained speech directed at device]
Level 3 - No Understanding:
User: [Inaudible or heavily accented speech]
System: "I didn't understand. Would you like to try again,
or I can turn on the bedroom lights for you?"
[Offers most common action as suggestion]
Level 4 - Repeated Failures:
After 3 failed attempts in 2 minutes:
System: "I'm having trouble hearing you today.
The light switch by the door also works,
or I can call for assistance."
[Graceful escalation to human help option]
Memory Support:
- Never require remembering exact syntax
- Offer suggestions proactively: “You can say things like ‘too cold’ or ‘lights brighter’”
- Recent commands available: “Do you want me to do the same as before?”
Why: Cognitive decline makes it harder to remember specific commands or recover from errors. Each failure increases frustration and decreases confidence. The system takes responsibility for misunderstanding rather than implying user error (“I didn’t understand” instead of “Invalid command”).
What we do: Implement room awareness so users do not need to specify location for every command.
Context Detection:
| User location |
Motion sensors, voice direction |
“Lights on” controls nearest room |
| Time of day |
Clock + learned patterns |
Morning = bedroom, Evening = living room |
| Recent activity |
Last room interacted with |
“A little warmer” adjusts same room as previous |
| Explicit override |
User says room name |
“Kitchen lights” overrides auto-detection |
Conversation Flow:
User: "Lights on"
[System detects user in living room via motion sensor]
System: "Living room lights are on."
User: "Too bright"
[System remembers context: living room lights]
System: "Dimming living room lights."
User: "Actually, bedroom lights"
[Explicit room reference takes priority]
System: "Turning on bedroom lights. Should I adjust living room too?"
Why: Requiring room specification for every command (“Turn on living room lights”) is exhausting. Natural conversation assumes context. However, the system announces which room it is affecting to prevent surprises – an elderly user in the living room should not wonder why bedroom lights came on.
What we do: Ensure core functions remain accessible when voice fails, respecting that no single modality works 100% of the time.
Multi-Modal Fallback Design:
| Voice command |
Physical wall switch |
Large button remote |
Caregiver call |
| “Lights on” |
Press illuminated button |
Tap bright-colored remote |
Button calls front desk |
| “Too cold” |
Thermostat dial (60pt numbers) |
Remote up/down buttons |
Staff notification |
Physical Control Requirements:
- Switches at 44 inches height (within ADA range of 15-48 inches for side approach)
- Large toggle switches (not small buttons)
- High contrast labeling (white on dark blue)
- Illuminated when off (findable in dark)
- Work during power outages (battery backup)
Remote Design:
- 5 large buttons only: Lights On, Lights Off, Warmer, Cooler, Help
- Tactile differentiation (bumps on Warmer, ridges on Cooler)
- Bright orange “Help” button always visible
- Weekly battery check notification to staff
Why: Voice-first does not mean voice-only. Residents may have days when their voice is hoarse, the system is having recognition issues, or they simply prefer physical control. Dignity means having options.
Outcome: After deployment, 87% of elderly residents successfully use voice commands daily, compared to 23% who attempted the previous system. Support calls for “it doesn’t understand me” dropped by 92%. Resident satisfaction surveys show 4.4/5 for ease of use.
Key Decisions Made:
| Intent-based NLU (not keywords) |
Supports natural speech patterns like “I’m cold” |
| Lower frequency audio responses |
Accommodates age-related high-frequency hearing loss |
| Maximum 2 choices in clarification |
Reduces cognitive load for memory-impaired users |
| System takes blame for errors |
Maintains user confidence (“I didn’t understand” not “Invalid command”) |
| Automatic room detection |
Eliminates need to remember/specify location each time |
| Physical switches remain primary |
Voice augments rather than replaces proven accessibility |
| Large, simple remote as backup |
Independent control when voice is not working |
| “Help” button always available |
Safety net for any situation |
Validation Method: Conduct in-home testing with 20 residents across hearing ability, cognitive status, and tech comfort levels. Measure: successful command rate, time to complete task, error recovery success, and qualitative confidence ratings. Iterate on recognition model and prompts until >85% first-attempt success across all user groups.