A real-time multimodal sentiment analysis system that combines facial expressions, environmental context, and speech to provide comprehensive emotional insights and recommendations.
SoothSayer is an intelligent assistant that analyzes your emotional state and surroundings in real-time using:
- Facial Expression Analysis - Detects emotions from facial cues and body language
- Environmental Context - Understands your surroundings and potential obstacles
- Speech Analysis - Transcribes and analyzes your spoken words for sentiment
- Movement Guidance - Calculates optimal movement angles for navigation
The system provides conversational feedback and recommendations based on this comprehensive analysis.
- Real-time Audio Recording - Continuous audio capture with configurable sensitivity
- Dual Camera System - Simultaneous front and back camera capture
- Beautiful UI - Modern, intuitive interface with real-time feedback
- Interval Recording - Automated recording and analysis cycles
- GROQ AI Integration - Uses state-of-the-art LLMs for analysis
- MiDaS Depth Estimation - 3D depth mapping for movement guidance
- Multimodal Processing - Combines visual, audio, and environmental data
- RESTful API - Clean endpoints for frontend communication
- Continuous Monitoring - Records and analyzes data every 10 seconds
- Instant Feedback - Provides immediate emotional insights
- Context Awareness - Understands both personal and environmental factors
- Facial Sentiment Analysis - Detects primary and secondary emotions
- Environmental Understanding - Identifies objects and context in surroundings
- Speech Transcription - Converts audio to text for content analysis
- Movement Optimization - Calculates optimal navigation paths
- Emotional Insights - Provides detailed emotional state analysis
- Conversational Responses - Natural language feedback and suggestions
- Contextual Guidance - Recommendations based on current situation
[Screenshots would be added here]
- React Native - Cross-platform mobile development
- Expo - Development platform and tools
- Expo Camera - Camera and audio recording
- Expo File System - File management and uploads
- Python 3.12+ - Core programming language
- Flask - Web framework for API
- GROQ API - AI/LLM services
- MiDaS - Depth estimation models
- OpenCV - Computer vision processing
- PyTorch - Deep learning framework
- Poetry - Dependency management
- GROQ Llama Models - Text analysis and generation
- GROQ Whisper - Speech-to-text transcription
- GROQ Vision Models - Image analysis and sentiment detection
- MiDaS - Monocular depth estimation
- Node.js 18+ and npm or yarn
- Python 3.12+
- Poetry (Python package manager)
- Expo CLI
- iOS Simulator or Android Emulator (for mobile testing)
- GROQ API Key - Get one at groq.com
git clone https://github.com/yourusername/soothsayer.git
cd soothsayercd backend
# Install Python dependencies
poetry install
# Set up environment variables
cp .env.example .env
# Edit .env and add your GROQ_API_KEY
# Run the backend server
poetry run python app.pyThe backend will start on http://localhost:5001
cd frontend
# Install dependencies
npm install
# Start the development server
npx expo start- Install Expo Go app on your mobile device
- Scan the QR code from the Expo development server
- Or run on iOS Simulator/Android Emulator
Create a .env file in the backend directory:
GROQ_API_KEY=your_groq_api_key_here-
Start Backend Server
cd backend poetry run python app.py -
Start Frontend Development Server
cd frontend npx expo start -
Open Mobile App
- Use Expo Go app to scan QR code
- Or press
ifor iOS Simulator /afor Android Emulator
-
Grant Permissions
- Allow microphone access for audio recording
- Allow camera access for photo capture
-
Configure Sensitivity
- Adjust the sensitivity slider (0-100%)
- Higher sensitivity = more frequent recordings
-
Start Recording
- Tap "Start Recording" to begin automated analysis
- The system will record every 10 seconds
- Photos are captured from both cameras simultaneously
-
View Results
- Analysis results appear in real-time
- Emotional insights and recommendations are provided
- Audio responses are generated and played
POST /api/analyze/face-sentiment- Analyze facial expressionsPOST /api/analyze/environment-sentiment- Analyze surroundingsPOST /api/analyze/audio-transcription- Transcribe speechPOST /api/analyze/combined-sentiment- Comprehensive multimodal analysis
POST /api/audio/upload- Upload audio filesPOST /api/photo/upload- Upload photosGET /api/audio/latest- Get latest audio fileGET /api/photo/latest- Get latest photos
GET /api/health- Health check
cd backend
poetry run python -c "from SoothSayer import SoothSayer; print('โ
Backend ready')"cd frontend
npm test-
GROQ API Errors
- Verify your API key is correct
- Check your GROQ account balance
- Ensure you're using the correct model names
-
Camera Permission Issues
- Grant camera permissions in device settings
- Restart the Expo development server
-
Audio Recording Problems
- Check microphone permissions
- Ensure device is not in silent mode
- Restart the app if needed
-
Backend Connection Issues
- Verify backend is running on port 5001
- Check firewall settings
- Ensure correct API endpoints in frontend
Enable detailed logging by setting environment variables:
export DEBUG=true
export LOG_LEVEL=DEBUG- Fork the repository
- Create a feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'Add amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Open a Pull Request
This project is licensed under the MIT License - see the LICENSE file for details.
- GROQ for providing fast AI inference
- Intel ISL for the MiDaS depth estimation models
- Expo for the excellent development platform
- React Native community for the robust mobile framework
- Issues: GitHub Issues
- Discussions: GitHub Discussions
- Email: your.email@example.com
- Voice Emotion Analysis - Analyze tone and pitch for emotional context
- Real-time Video Analysis - Continuous video sentiment analysis
- Machine Learning Models - Custom trained models for specific use cases
- Cloud Deployment - AWS/Azure deployment options
- Mobile App Stores - iOS App Store and Google Play Store releases
- Web Dashboard - Browser-based analysis interface
- API Documentation - Interactive API documentation with Swagger
- Multi-language Support - Internationalization for global users
Made with โค๏ธ by the SoothSayer Team