Fix prompt, model, and temperature overrides in embed chat widget (#4036)
fix data-prompt attribute in embed chat widget Co-authored-by: Timothy Carambat <rambat1010@gmail.com>
This commit is contained in:
parent
b7773f2f9b
commit
62d3803bf0
@ -39,9 +39,9 @@ function embeddedEndpoints(app) {
|
||||
response.flushHeaders();
|
||||
|
||||
await streamChatWithForEmbed(response, embed, message, sessionId, {
|
||||
prompt,
|
||||
model,
|
||||
temperature,
|
||||
promptOverride: prompt,
|
||||
modelOverride: model,
|
||||
temperatureOverride: temperature,
|
||||
username,
|
||||
});
|
||||
await Telemetry.sendTelemetry("embed_sent_chat", {
|
||||
|
||||
Loading…
Reference in New Issue
Block a user