Opsmate - DataOps JIRA Agent
Inspiration
We watched a healthcare DataOps PM spend 3 hours every morning creating JIRA tickets from email requests - many nearly identical to last week's, but requiring manual processing because patient data couldn't leave their servers.
The problem is universal for data teams:
- 70% unplanned work scattered across emails, Slack, and meetings
- 30% of their day lost to repetitive ticket creation
- Zero automation options for teams handling PII/PHI data
- No learning from patterns - same requests processed manually every time
They needed automation that was intelligent yet completely private.
What it does
Opsmate automatically transforms unstructured requests from emails, Slack messages, and documentation into ready-to-approve JIRA tickets in under 60 seconds.
Core capabilities:
- Multi-source extraction: Monitors emails, Slack, Teams, and file drops
- Local LLM processing: Uses Ollama-hosted models, keeping data on-premise
- Intelligent scraping: Browser Use extracts knowledge from Confluence/wikis
- PII/PHI protection: Automatic detection and masking for compliance
- Pattern learning: Identifies recurring requests and applies templates
- Human-in-the-loop: Full editing control before ticket creation
The system builds a growing knowledge base with every interaction, improving from 70% to 92% accuracy within weeks.
How we built it
We started by deploying Ollama locally to host Llama2 for text extraction without external API calls. Our Python backend monitors email inboxes (IMAP) and Slack channels (webhooks), feeding content through our extraction pipeline.
Technical implementation:
- FastAPI backend for REST APIs and WebSocket connections
- Browser Use integration for scraping JavaScript-heavy documentation sites
- PostgreSQL + Qdrant for hybrid structured/vector storage
- React frontend with Material-UI for the review interface
- Docker containers for easy deployment and scaling
- Redis for caching and real-time state management
Browser Use became our secret weapon - capturing documentation context from internal wikis that lacked APIs, dramatically improving extraction accuracy.
Challenges we ran into
Dynamic content extraction was our biggest hurdle. Internal documentation sites use heavy JavaScript rendering that traditional scrapers couldn't handle. Browser Use solved this by providing full browser automation capabilities.
Other key challenges:
- PII detection balance: Too aggressive and we lost context, too lenient risked compliance
- LLM hallucinations: Model invented plausible table names - fixed with schema validation
- User trust: Initial "fully automated" approach was rejected - PMs needed control
- Performance at scale: Local LLMs + hundreds of requests required careful optimization
- Session management: Browser Use sessions for efficient repeated scraping
Accomplishments that we're proud of
We reduced ticket creation time from 10+ minutes to under 60 seconds while maintaining 100% data privacy - no information ever leaves the organization's servers.
Key achievements:
- 92% extraction accuracy through Browser Use knowledge scraping
- 100% PII/PHI compliance with automatic detection and masking
- 15 hours saved weekly per PM on ticket creation
- Pattern library that learns and improves with every correction
- Zero external API calls - complete on-premise operation
- Trust through transparency - full visibility and control for users
What we learned
Browser automation with Browser Use was transformative - 80% of organizational knowledge lives in wikis and documentation sites without APIs. This integration alone improved our accuracy by 20%.
Key insights:
- Local LLMs are sufficient for extraction tasks - no need for GPT-4
- Simple patterns beat complex AI for 80% of use cases
- Human control builds trust - review and edit capabilities are essential
- Context is everything - similar past tickets improved accuracy dramatically
- Privacy-first architecture is non-negotiable for enterprise adoption
What's next for Opsmate
We're expanding Browser Use integration to scrape SharePoint, internal knowledge bases, and legacy systems that trap valuable context. The ability to interact with any web-based system opens unlimited possibilities.
Roadmap priorities:
- Voice input for capturing requests directly from meetings
- Shareable pattern library for collective intelligence across teams
- Fine-tuned models specifically for ticket extraction
- Predictive analytics to forecast request volumes
- Multi-language support for global teams
- Advanced Browser Use features: Form filling, multi-tab workflows, and complex navigation patterns
Our vision is to free DataOps teams from operational overhead, letting them focus on delivering value rather than managing tickets.
Built With
- argon2
- beautiful-soup
- browser-use
- celery
- confluence-api
- docker
- docker-compose
- eslint
- fastapi
- imap
- javascript
- jest
- jira-rest-api
- jwt
- llama2
- material-ui
- microsoft-teams-webhooks
- minio
- mistral
- next
- nginx
- node.js
- ollama
- openapi
- playwright
- poetry
- postgresql
- prettier
- pytest
- python
- qdrant
- rabbitmq
- react
- react-query
- recharts
- redis
- redux-toolkit
- scikit-learn
- sentence-transformers
- slack-api
- smtp
- socket.io
- spacy
- swagger
- typescript
Log in or sign up for Devpost to join the conversation.