Conversation
🦋 Changeset detectedLatest commit: ea43888 The changes in this PR will be included in the next version bump. This PR includes changesets to release 1 package
Not sure what this means? Click here to learn what changesets are. Click here if you're a maintainer who wants to add another changeset to this PR |
| elif "contentBlockStop" in chunk: | ||
| if self._text: | ||
| # ignore chain of thought | ||
| text = re.sub(r"<thinking>.*?</thinking>", "", self._text) |
There was a problem hiding this comment.
should this be AWS specific? or a general thing we should do in the LLM superclass? maybe an option: remove_thinking?
also is it guaranteed the chain of thought appears in the same chunk?
There was a problem hiding this comment.
I think we should keep the thinking tokens (in v1.0 at least).
It should be the responsability of the tts_node to remove those tokens.
Tho not against merging this PR for main as a temporary solution
There was a problem hiding this comment.
we need a util to strip tags in a streaming fashion
fixes: #1599