forked from nlweb-ai/NLWeb
-
Notifications
You must be signed in to change notification settings - Fork 0
Customizations #1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Draft
schmichri
wants to merge
28
commits into
main
Choose a base branch
from
iuneracustomizations
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Draft
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Update OpenAI embedding chunk size to 4000 tokens (previously 6000)
# Update prompts.py with new functionality for compact representation of final_ranked_answers feat(prompts): Extract essential fields from final_ranked_answers and create a compact representation for use in prompts. ```
… with sorting by modificationDate`
This commit message follows the style of recent commits by prioritizing the main action, using a concise description, and avoiding unnecessary details. It also adheres to Markdown formatting conventions and includes the programming language (`Python`) in the code block.
Squash merge Docker branch
This commit message follows the style of recent commits by prioritizing the main action, using a concise description, and avoiding unnecessary details. It also adheres to Markdown formatting conventions and includes the programming language (`
commit 65d588d Merge: 3560a1f 60e98e8 Author: Chelsea <chelseacarter29@users.noreply.github.com> Date: Thu Jun 12 16:28:23 2025 -0700 Merge pull request nlweb-ai#194 from microsoft/Scraper Basic scraper commit 60e98e8 Author: R.V.Guha <guha@guha.com> Date: Thu Jun 12 16:08:49 2025 -0700 Incremental crawl fixing Incremental Crawl Fixing commit 0e676cc Author: R.V.Guha <guha@guha.com> Date: Thu Jun 12 14:55:28 2025 -0700 incremental crawl Incremental crawl commit 18d1169 Author: R.V.Guha <guha@guha.com> Date: Thu Jun 12 09:32:35 2025 -0700 Create README.md commit be325c1 Author: R.V.Guha <guha@guha.com> Date: Wed Jun 11 09:29:14 2025 -0700 Basic scraper Read the sitemps to get the URLs, get the site content, extract the markup, compute embeddings, load into db # Conflicts: # .gitignore
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Pull Request Summary
This pull request introduces a new, powerful feature for analyzing and responding to user queries using a Large Language Model (LLM), alongside significant enhancements to the CI/CD pipeline. Additionally, it addresses the need to restore critical performance features that were not present in the initial branch.
New Features & Enhancements
LLM-Powered Query Analysis:
analyze_query.py``prompt_runner.pyReal-time Streaming Support:
StreamingWrapper.pyCI/CD Automation:
prod-build.ymlDockerfilein themainoriuneracustomizationsbranches, the workflow will automatically extract the version, build the image, push it to the registry, and create a corresponding Git tag.Dockerfile Improvements:
Dockerfileis now configured for non-interactive builds and the applicationVERSIONhas been updated to2.9.0.Reinstatement of Performance Features
It was noted that crucial performance-related features were missing from this branch. This PR will also focus on restoring the following functionalities to ensure the application remains fast and efficient:
By combining the new LLM capabilities with these critical performance optimizations, this pull request delivers a significantly more powerful and efficient application.