Merge branch 'main' into hosted
This commit is contained in:
commit
f9d6279b53
@ -89,7 +89,7 @@ async def stream_claude_response(
|
|||||||
client = AsyncAnthropic(api_key=api_key)
|
client = AsyncAnthropic(api_key=api_key)
|
||||||
|
|
||||||
# Base parameters
|
# Base parameters
|
||||||
max_tokens = 4096
|
max_tokens = 8192
|
||||||
temperature = 0.0
|
temperature = 0.0
|
||||||
|
|
||||||
# Translate OpenAI messages to Claude messages
|
# Translate OpenAI messages to Claude messages
|
||||||
@ -127,6 +127,7 @@ async def stream_claude_response(
|
|||||||
temperature=temperature,
|
temperature=temperature,
|
||||||
system=system_prompt,
|
system=system_prompt,
|
||||||
messages=claude_messages, # type: ignore
|
messages=claude_messages, # type: ignore
|
||||||
|
extra_headers={"anthropic-beta": "max-tokens-3-5-sonnet-2024-07-15"},
|
||||||
) as stream:
|
) as stream:
|
||||||
async for text in stream.text_stream:
|
async for text in stream.text_stream:
|
||||||
await callback(text)
|
await callback(text)
|
||||||
|
|||||||
@ -13,8 +13,8 @@ from evals.config import EVALS_DIR
|
|||||||
from evals.core import generate_code_core
|
from evals.core import generate_code_core
|
||||||
from evals.utils import image_to_data_url
|
from evals.utils import image_to_data_url
|
||||||
|
|
||||||
STACK = "ionic_tailwind"
|
STACK = "html_tailwind"
|
||||||
MODEL = Llm.GPT_4O_2024_05_13
|
MODEL = Llm.CLAUDE_3_5_SONNET_2024_06_20
|
||||||
N = 1 # Number of outputs to generate
|
N = 1 # Number of outputs to generate
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@ -16,14 +16,16 @@ describe("e2e tests", () => {
|
|||||||
let browser: Browser;
|
let browser: Browser;
|
||||||
let page: Page;
|
let page: Page;
|
||||||
|
|
||||||
const DEBUG = false;
|
const DEBUG = true;
|
||||||
const IS_HEADLESS = true;
|
const IS_HEADLESS = true;
|
||||||
|
|
||||||
const stacks = Object.values(Stack).slice(0, DEBUG ? 1 : undefined);
|
const stacks = Object.values(Stack).slice(0, DEBUG ? 1 : undefined);
|
||||||
const models = Object.values(CodeGenerationModel).slice(
|
const models = DEBUG
|
||||||
0,
|
? [
|
||||||
DEBUG ? 1 : undefined
|
CodeGenerationModel.GPT_4O_2024_05_13,
|
||||||
);
|
CodeGenerationModel.CLAUDE_3_5_SONNET_2024_06_20,
|
||||||
|
]
|
||||||
|
: Object.values(CodeGenerationModel);
|
||||||
|
|
||||||
beforeAll(async () => {
|
beforeAll(async () => {
|
||||||
browser = await puppeteer.launch({ headless: IS_HEADLESS });
|
browser = await puppeteer.launch({ headless: IS_HEADLESS });
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user