Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
9 changes: 7 additions & 2 deletions commands/dedicated_inference_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -23,8 +23,9 @@ var (
EnablePublicEndpoint: true,
ModelDeployments: []*godo.DedicatedInferenceModelRequest{
{
ModelSlug: "mistral/mistral-7b-instruct-v3",
ModelProvider: "hugging_face",
ModelSlug: "mistral/mistral-7b-instruct-v3",
ModelProvider: "hugging_face",
ProviderModelID: "mistralai/Mistral-7B-Instruct-v0.3",
Accelerators: []*godo.DedicatedInferenceAcceleratorRequest{
{
Scale: 2,
Expand Down Expand Up @@ -96,6 +97,7 @@ func TestRunDedicatedInferenceCreate(t *testing.T) {
"model_deployments": [
{
"model_slug": "mistral/mistral-7b-instruct-v3",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_provider": "hugging_face",
"accelerators": [
{"scale": 2, "type": "prefill", "accelerator_slug": "gpu-mi300x1-192gb"},
Expand Down Expand Up @@ -132,6 +134,7 @@ func TestRunDedicatedInferenceCreate_WithHuggingFaceToken(t *testing.T) {
"model_deployments": [
{
"model_slug": "mistral/mistral-7b-instruct-v3",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_provider": "hugging_face",
"accelerators": [
{"scale": 2, "type": "prefill", "accelerator_slug": "gpu-mi300x1-192gb"},
Expand Down Expand Up @@ -202,6 +205,7 @@ func TestRunDedicatedInferenceUpdate(t *testing.T) {
"model_deployments": [
{
"model_slug": "mistral/mistral-7b-instruct-v3",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_provider": "hugging_face",
"accelerators": [
{"scale": 2, "type": "prefill", "accelerator_slug": "gpu-mi300x1-192gb"},
Expand Down Expand Up @@ -239,6 +243,7 @@ func TestRunDedicatedInferenceUpdate_WithHuggingFaceToken(t *testing.T) {
"model_deployments": [
{
"model_slug": "mistral/mistral-7b-instruct-v3",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_provider": "hugging_face",
"accelerators": [
{"scale": 2, "type": "prefill", "accelerator_slug": "gpu-mi300x1-192gb"},
Expand Down
2 changes: 1 addition & 1 deletion go.mod
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ go 1.25.0
require (
github.com/blang/semver v3.5.1+incompatible
github.com/creack/pty v1.1.21
github.com/digitalocean/godo v1.184.0
github.com/digitalocean/godo v1.184.1-0.20260410132851-6a015e2e3b32
github.com/docker/cli v24.0.5+incompatible
github.com/docker/docker v25.0.6+incompatible
github.com/docker/docker-credential-helpers v0.7.0 // indirect
Expand Down
4 changes: 2 additions & 2 deletions go.sum
Original file line number Diff line number Diff line change
Expand Up @@ -91,8 +91,8 @@ github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSs
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc h1:U9qPSI2PIWSS1VwoXQT9A3Wy9MM3WgvqSxFWenqJduM=
github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
github.com/digitalocean/godo v1.184.0 h1:2B2CQhxftlf3xa24Nrzn5CBQlaQjyaWqi3XbbnJlG3w=
github.com/digitalocean/godo v1.184.0/go.mod h1:xQsWpVCCbkDrWisHA72hPzPlnC+4W5w/McZY5ij9uvU=
github.com/digitalocean/godo v1.184.1-0.20260410132851-6a015e2e3b32 h1:bO0M6ODWgllf/LlR0wzlxIxiMMcJuHvlrQdDfhC4ufY=
github.com/digitalocean/godo v1.184.1-0.20260410132851-6a015e2e3b32/go.mod h1:xQsWpVCCbkDrWisHA72hPzPlnC+4W5w/McZY5ij9uvU=
github.com/distribution/reference v0.6.0 h1:0IXCQ5g4/QMHHkarYzh5l+u8T3t73zM5QvfrDyIgxBk=
github.com/distribution/reference v0.6.0/go.mod h1:BbU0aIcezP1/5jX/8MP0YiH4SdvB5Y4f/wlDRiLyi3E=
github.com/docker/cli v24.0.5+incompatible h1:WeBimjvS0eKdH4Ygx+ihVq1Q++xg36M/rMi4aXAvodc=
Expand Down
2 changes: 2 additions & 0 deletions integration/dedicated_inference_get_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -183,6 +183,7 @@ ID Name Status
"model_deployments": [
{
"model_id": "model-001",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_slug": "mistral/mistral-7b-instruct-v3",
"model_provider": "hugging_face",
"accelerators": [
Expand Down Expand Up @@ -218,6 +219,7 @@ ID Name Status
"model_deployments": [
{
"model_id": "model-001",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_slug": "mistral/mistral-7b-instruct-v3",
"model_provider": "hugging_face",
"accelerators": [
Expand Down
4 changes: 4 additions & 0 deletions integration/dedicated_inference_list_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -176,6 +176,7 @@ ID Name Status
"name": "test-di-1",
"region": "nyc2",
"status": "ACTIVE",
"provider_model_id": ["mistralai/Mistral-7B-Instruct-v0.3"],
"vpc_uuid": "00000000-0000-4000-8000-000000000001",
"endpoints": {
"public_endpoint_fqdn": "public.di-1.example.com",
Expand All @@ -189,6 +190,7 @@ ID Name Status
"name": "test-di-2",
"region": "sfo3",
"status": "PROVISIONING",
"provider_model_id": ["meta-llama/Meta-Llama-3-8B-Instruct"],
"vpc_uuid": "11111111-1111-4111-8111-111111111112",
"endpoints": {
"public_endpoint_fqdn": "public.di-2.example.com",
Expand All @@ -212,6 +214,7 @@ ID Name Status
"name": "test-di-1",
"region": "nyc2",
"status": "ACTIVE",
"provider_model_id": ["mistralai/Mistral-7B-Instruct-v0.3"],
"vpc_uuid": "00000000-0000-4000-8000-000000000001",
"endpoints": {
"public_endpoint_fqdn": "public.di-1.example.com",
Expand All @@ -235,6 +238,7 @@ ID Name Status
"name": "test-di-1",
"region": "nyc2",
"status": "ACTIVE",
"provider_model_id": ["mistralai/Mistral-7B-Instruct-v0.3"],
"vpc_uuid": "00000000-0000-4000-8000-000000000001",
"endpoints": {
"public_endpoint_fqdn": "public.di-1.example.com",
Expand Down
5 changes: 4 additions & 1 deletion integration/dedicated_inference_update_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -270,7 +270,8 @@ ID Name Status
"enable_public_endpoint": true,
"model_deployments": [
{
"model_slug": "mistral/mistral-7b-instruct-v3",
"model_slug": "mistral/mistral-7b-instruct-v3",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_provider": "hugging_face",
"accelerators": [
{"scale": 3, "type": "prefill", "accelerator_slug": "gpu-mi300x1-192gb"},
Expand All @@ -288,6 +289,7 @@ vpc:
enable_public_endpoint: true
model_deployments:
- model_slug: mistral/mistral-7b-instruct-v3
provider_model_id: mistralai/Mistral-7B-Instruct-v0.3
model_provider: hugging_face
accelerators:
- scale: 3
Expand Down Expand Up @@ -322,6 +324,7 @@ model_deployments:
"model_deployments": [
{
"model_id": "model-001",
"provider_model_id": "mistralai/Mistral-7B-Instruct-v0.3",
"model_slug": "mistral/mistral-7b-instruct-v3",
"model_provider": "hugging_face",
"accelerators": [
Expand Down
37 changes: 20 additions & 17 deletions vendor/github.com/digitalocean/godo/dedicated_inference.go

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

2 changes: 1 addition & 1 deletion vendor/modules.txt
Original file line number Diff line number Diff line change
Expand Up @@ -61,7 +61,7 @@ github.com/creack/pty
# github.com/davecgh/go-spew v1.1.2-0.20180830191138-d8f796af33cc
## explicit
github.com/davecgh/go-spew/spew
# github.com/digitalocean/godo v1.184.0
# github.com/digitalocean/godo v1.184.1-0.20260410132851-6a015e2e3b32
## explicit; go 1.23.0
github.com/digitalocean/godo
github.com/digitalocean/godo/metrics
Expand Down
Loading