Inspiration
During emergencies, every second counts. We wanted to create an AI-powered system that allows people to communicate in any language, even in high-stress or physically restrictive situations. Whether someone’s trapped or multitasking, Res-Q makes it possible to get help or access critical info using only your voice.
Our goal was to combine real-time speech interaction, AI reasoning with multilingual capabilities, and natural voice synthesis to build a tool that’s both practical and human-friendly, not meant to replace police or services but rather to supplement them.
What it does
Res-Q is a voice-based AI assistant that enables natural, conversational interaction with an emergency response system. With just your voice, you can:
- Ask for help or describe a situation
- Get real-time info scraped from trusted sources (e.g., weather alerts, nearby hospitals)
- Automatically detect your approximate location using IP-based geolocation
- Hear natural, lifelike voice responses generated by ElevenLabs
- Have intelligent, context-aware conversations powered by Google Gemini AI
And the best part is, you can do this in any language!
How we built it
Res-Q is built on top of the Gemini Voice Base Stack, integrating:
- SpeechRecognition + PyAudio for continuous microphone listening
- Google Gemini API for text generation and reasoning
- ElevenLabs API for real-time text-to-speech (TTS)
- IP-based geolocation for automatic context detection
- Web scraping to pull in real-time information during conversations
- Python backend with a React.js frontend
Challenges we ran into
- Integrating frontend and backend
- Multilingual aspect
- Managing real-time audio streaming without lag
- Syncing speech detection and TTS playback
- Handling API latency for smooth interaction
Accomplishments that we're proud of
- Built a fully functional real-time voice AI stack from scratch
- Achieved smooth voice interaction with Gemini and ElevenLabs
- Implemented auto-pause detection and interruptible playback
- Designed the foundation for a rescue-focused AI assistant
What we learned
- How to integrate LLMs with audio interfaces
- Optimizing real-time speech recognition pipelines
- Handling multi-threaded playback and recording in Python
- Designing for hands-free accessibility and inclusivity
What's next for Res-Q
- Launch a web dashboard for monitoring and control
- Create a mobile version for on-the-go use
- Make more flexible with different languages
Log in or sign up for Devpost to join the conversation.