Skip to content

Commit

Permalink
Replace Cerebras Llama 3.1 70b with Llama 3.3 70b (#4743)
Browse files Browse the repository at this point in the history
  • Loading branch information
henrytwo authored Dec 17, 2024
1 parent 63fe24d commit 960fbf0
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 7 deletions.
2 changes: 1 addition & 1 deletion autogen/oai/cerebras.py
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@
CEREBRAS_PRICING_1K = {
# Convert pricing per million to per thousand tokens.
"llama3.1-8b": (0.10 / 1000, 0.10 / 1000),
"llama3.1-70b": (0.60 / 1000, 0.60 / 1000),
"llama-3.3-70b": (0.85 / 1000, 1.20 / 1000),
}


Expand Down
12 changes: 6 additions & 6 deletions test/oai/test_cerebras.py
Original file line number Diff line number Diff line change
Expand Up @@ -136,7 +136,7 @@ def test_cost_calculation(mock_response):
choices=[{"message": "Test message 1"}],
usage={"prompt_tokens": 500, "completion_tokens": 300, "total_tokens": 800},
cost=None,
model="llama3.1-70b",
model="llama-3.3-70b",
)
calculated_cost = calculate_cerebras_cost(
response.usage["prompt_tokens"], response.usage["completion_tokens"], response.model
Expand All @@ -160,15 +160,15 @@ def test_create_response(mock_chat, cerebras_client):
MagicMock(finish_reason="stop", message=MagicMock(content="Example Cerebras response", tool_calls=None))
]
mock_cerebras_response.id = "mock_cerebras_response_id"
mock_cerebras_response.model = "llama3.1-70b"
mock_cerebras_response.model = "llama-3.3-70b"
mock_cerebras_response.usage = MagicMock(prompt_tokens=10, completion_tokens=20) # Example token usage

mock_chat.return_value = mock_cerebras_response

# Test parameters
params = {
"messages": [{"role": "user", "content": "Hello"}, {"role": "assistant", "content": "World"}],
"model": "llama3.1-70b",
"model": "llama-3.3-70b",
}

# Call the create method
Expand All @@ -179,7 +179,7 @@ def test_create_response(mock_chat, cerebras_client):
response.choices[0].message.content == "Example Cerebras response"
), "Response content should match expected output"
assert response.id == "mock_cerebras_response_id", "Response ID should match the mocked response ID"
assert response.model == "llama3.1-70b", "Response model should match the mocked response model"
assert response.model == "llama-3.3-70b", "Response model should match the mocked response model"
assert response.usage.prompt_tokens == 10, "Response prompt tokens should match the mocked response usage"
assert response.usage.completion_tokens == 20, "Response completion tokens should match the mocked response usage"

Expand Down Expand Up @@ -211,7 +211,7 @@ def test_create_response_with_tool_call(mock_chat, cerebras_client):
)
],
id="mock_cerebras_response_id",
model="llama3.1-70b",
model="llama-3.3-70b",
usage=MagicMock(prompt_tokens=10, completion_tokens=20),
)

Expand Down Expand Up @@ -239,7 +239,7 @@ def test_create_response_with_tool_call(mock_chat, cerebras_client):

# Call the create method
response = cerebras_client.create(
{"messages": cerebras_messages, "tools": converted_functions, "model": "llama3.1-70b"}
{"messages": cerebras_messages, "tools": converted_functions, "model": "llama-3.3-70b"}
)

# Assertions to check if the functions and content are included in the response
Expand Down

0 comments on commit 960fbf0

Please sign in to comment.