-
-
Notifications
You must be signed in to change notification settings - Fork 10
Closed
Labels
enhancementNew feature or requestNew feature or requesthelp wantedExtra attention is neededExtra attention is needed
Description
This proved a bit tricky, because the MLC library works based on a callback mechanism:
from mlc_chat import ChatModule
from mlc_chat.callback import StreamToStdout
cm = ChatModule(model="Llama-2-7b-chat-hf-q4f16_1")
cm.generate(
prompt="A poem about a bunny eating lunch",
progress_callback=StreamToStdout(callback_interval=1),
)But... LLM expects to be able to do something like this:
for chunk in cm.generate(...):
yield chunkReactions are currently unavailable
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or requesthelp wantedExtra attention is neededExtra attention is needed