Imagine you're lost and alone in a remote environment, you have no internet access to call for help or look up survival knowledge and time is running out.
Whether it's a hiker with a broken leg, a soldier in combat, a responder during a natural disaster, or someone lost at sea, the lack of internet connectivity and immediate expert guidance can mean the difference between life and death.
SurviveX is an offline-first embedded AI assistant that combines Edge AI on a hands free device like the Apple Vision Pro along with health monitoring and voice guidance to provide real-time survival assistance.
- Our Assistant that provides step-by-step guidance, helpful for providing stressed users with fast responses and ample opportunity for clarification throughout the process all in the absence of an internet connection.
- Tracking vital signs through Terra's API, and provide step-by-step voice guidance.
- Our Machine Learning Model is designed to run on a hands free device like the Apple Vision Pro enabling survivors, first responders and soldiers to communicate via speech so your hands can focus on what matters : survival.
- Fine tuning our Model to provide a user specific solution governed by environment indicative stress of the situation that users may be in.
Whether you need to treat an injury, start a fire, fix up your broken car on the side of the highway, or navigate by stars, SurvivalX acts as your personal survival expert, adapting its guidance based on your speech and environmental conditions.
- We used ExecuTorch for on device inference for our Edge AI solution.
- Note : Here is a link to our Google Drive containing the
.ptefiles generated to build and run our model. - The model we decided on was Meta's Llama-3.2-1B-Instruct since it was small and the most practical run on device.
- To fine tune our model we had to use torchtune - a PyTorch library for fine tuning our Llama Model on an instance of Nvidia's H-100 using Brev.dev.
- We used SwiftUI for the interface on VisionOS and Swift for implementing the on device Llama model.
- We used data from Terra API to stimulate tracking of heartbeats and vitals.
- Clone PyTorch ExecuTorch and it's submodules. No need to install.
- In
examples/demo-apps/apple_iosclone this repo. - In
SurviveX/SurviveX/Llama/addllaama3_2.pteandtokenizer.modelfrom our Google Drive. - Open in Xcode and build as specified here.