Inspiration

For over 8.7 million visually impaired adults in the U.S., something as routine as filling out a tax form, medical intake sheet, or job application can be frustrating or even impossible without help. Existing form fillers require visual interaction — clicking, dragging, typing — and none are built with true accessibility in mind. We wanted to change that by building a tool that gives these users independence, speed, and confidence.

What it does

fill.ai is a voice-powered, AI-driven form filler designed specifically for visually impaired users. Just upload any form — PDF, scan, or image — and the app:

  • Automatically detects fields using AI + OCR
  • Prompts the user to fill out each field using natural language
  • Allows users to speak their responses entirely by voice
  • Auto-fills the form in real time and generates a completed PDF

No mouse. No keyboard. No visual interface required.

How we built it

  • Frontend: React + Vite + SCSS Modules, with accessible markup and keyboard navigation support.
  • Voice Input: Web Speech API for speech-to-text conversion.
  • OCR & Field Detection: Tesseract.js + custom logic to parse text layout and detect form fields from scanned documents.
  • Form Filling Logic: JSON-based structure for field mapping, tied to voice prompts and AI suggestions.
  • PDF Handling: PDF-lib to generate and fill form data into PDF templates.

Challenges we ran into

  • OCR Accuracy: Scanned forms are often low-quality or skewed. We had to implement cleaning logic and fallback detection methods.
  • Voice Handling: Managing speech input in a structured and user-friendly way was tricky, especially with multiple fields and interruptions.
  • Form Complexity: Real-world forms are inconsistent — we had to account for variable layouts and missing field tags.

Accomplishments that we're proud of

  • Created a fully voice-driven form filling experience — no mouse or keyboard needed.
  • Built accessible UI components that work well with screen readers.
  • Successfully processed and completed real scanned forms using only voice input.
  • Designed the system to be useful not just for the visually impaired, but for anyone needing hands-free interaction.
  • Successfuly implemented language recognition for various languages including Hindi, Spanish, Ukrainian and Hurdu

What we learned

  • Accessibility-first design isn't just a feature — it changes how you think about user flows and interface priorities.
  • Voice UI is incredibly powerful, but needs thoughtful structure and fallback handling.
  • AI can enhance accessibility when it’s used with purpose — detecting form fields from imperfect scans was a real win.

What's next for fill.ai

  • 🔄 Improve field detection using ML-based layout analysis
  • 🌐 Expand language support for multilingual users
  • 📱 Build a mobile-first experience for on-the-go form filling
  • 🧑‍🦯 Partner with accessibility orgs for real user testing and feedback
  • 🔒 Add secure document upload and signing capabilities

Built With

Share this project:

Updates

posted an update

Announcement!!

Thanks to some last minute hacking we were able to get language translation working with Fill.AI, so you are able to communicate with PDF's in you native language while it fills the fields in English!!

Log in or sign up for Devpost to join the conversation.