Starting April 29, 2025, Gemini 1.5 Pro and Gemini 1.5 Flash models are not available in projects that have no prior usage of these models, including new projects. For details, see Model versions and lifecycle.
Stay organized with collections
Save and categorize content based on your preferences.
Qwen models on Vertex AI offer fully managed and serverless
models as APIs. To use a Qwen model on Vertex AI, send
a request directly to the Vertex AI API endpoint. Because
Qwen models use a managed API, there's no need to provision or
manage infrastructure.
You can stream your responses to reduce the end-user latency perception. A
streamed response uses server-sent events (SSE) to incrementally stream the
response.
Available Qwen models
The following models are available from Qwen to use in
Vertex AI. To access a Qwen model, go to its
Model Garden model card.
Qwen3 Coder (Qwen3 Coder)
Qwen3 Coder (Qwen3 Coder) is a large-scale, open-weight model
developed for advanced software development tasks. The model's key feature is
its large context window, allowing it to process and understand large codebases
comprehensively.
Qwen3 235B (Qwen3 235B) is a large 235B parameter model. The model
is distinguished by its "hybrid thinking" capability, which allows users to
dynamically switch between a methodical, step-by-step "thinking" mode for
complex tasks like mathematical reasoning and coding, and a rapid "non-thinking"
mode for general-purpose conversation. Its large context window makes it
suitable for use cases requiring deep reasoning and long-form comprehension.
To use Qwen models with Vertex AI, you must perform the
following steps. The Vertex AI API
(aiplatform.googleapis.com) must be enabled to use
Vertex AI. If you already have an existing project with the
Vertex AI API enabled, you can use that project instead of creating a
new project.
Sign in to your Google Cloud account. If you're new to
Google Cloud,
create an account to evaluate how our products perform in
real-world scenarios. New customers also get $300 in free credits to
run, test, and deploy workloads.
In the Google Cloud console, on the project selector page,
select or create a Google Cloud project.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Hard to understand","hardToUnderstand","thumb-down"],["Incorrect information or sample code","incorrectInformationOrSampleCode","thumb-down"],["Missing the information/samples I need","missingTheInformationSamplesINeed","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2025-08-27 UTC."],[],[],null,["# Qwen models\n\n| **Note:** Qwen models are not a Google product, and its availability in Vertex AI is subject to the terms for \"Separate Offerings\" in the AI/ML Services section of the [Service Specific\n| Terms](/terms/service-terms), and separate terms found in the relevant model card.\n\nQwen models on Vertex AI offer fully managed and serverless\nmodels as APIs. To use a Qwen model on Vertex AI, send\na request directly to the Vertex AI API endpoint. Because\nQwen models use a managed API, there's no need to provision or\nmanage infrastructure.\n\nYou can stream your responses to reduce the end-user latency perception. A\nstreamed response uses *server-sent events* (SSE) to incrementally stream the\nresponse.\n\nAvailable Qwen models\n---------------------\n\nThe following models are available from Qwen to use in\nVertex AI. To access a Qwen model, go to its\nModel Garden model card.\n\n### Qwen3 Coder (Qwen3 Coder)\n\nQwen3 Coder (`Qwen3 Coder`) is a large-scale, open-weight model\ndeveloped for advanced software development tasks. The model's key feature is\nits large context window, allowing it to process and understand large codebases\ncomprehensively.\n\n[Go to the Qwen3 Coder model card](https://console.cloud.google.com/vertex-ai/publishers/qwen/model-garden/qwen3-coder-480b-a35b-instruct-maas)\n\n### Qwen3 235B (Qwen3 235B)\n\nQwen3 235B (`Qwen3 235B`) is a large 235B parameter model. The model\nis distinguished by its \"hybrid thinking\" capability, which allows users to\ndynamically switch between a methodical, step-by-step \"thinking\" mode for\ncomplex tasks like mathematical reasoning and coding, and a rapid \"non-thinking\"\nmode for general-purpose conversation. Its large context window makes it\nsuitable for use cases requiring deep reasoning and long-form comprehension.\n\n[Go to the Qwen3 235B model card](https://console.cloud.google.com/vertex-ai/publishers/qwen/model-garden/qwen3-235b-a22b-instruct-2507-maas)\n\n### Before you begin\n\nTo use Qwen models with Vertex AI, you must perform the\nfollowing steps. The Vertex AI API\n(`aiplatform.googleapis.com`) must be enabled to use\nVertex AI. If you already have an existing project with the\nVertex AI API enabled, you can use that project instead of creating a\nnew project.\n\n- Sign in to your Google Cloud account. If you're new to Google Cloud, [create an account](https://console.cloud.google.com/freetrial) to evaluate how our products perform in real-world scenarios. New customers also get $300 in free credits to run, test, and deploy workloads.\n- In the Google Cloud console, on the project selector page,\n select or create a Google Cloud project.\n\n [Go to project selector](https://console.cloud.google.com/projectselector2/home/dashboard)\n-\n [Verify that billing is enabled for your Google Cloud project](/billing/docs/how-to/verify-billing-enabled#confirm_billing_is_enabled_on_a_project).\n\n-\n\n\n Enable the Vertex AI API.\n\n\n [Enable the API](https://console.cloud.google.com/flows/enableapi?apiid=aiplatform.googleapis.com)\n\n- In the Google Cloud console, on the project selector page,\n select or create a Google Cloud project.\n\n [Go to project selector](https://console.cloud.google.com/projectselector2/home/dashboard)\n-\n [Verify that billing is enabled for your Google Cloud project](/billing/docs/how-to/verify-billing-enabled#confirm_billing_is_enabled_on_a_project).\n\n-\n\n\n Enable the Vertex AI API.\n\n\n [Enable the API](https://console.cloud.google.com/flows/enableapi?apiid=aiplatform.googleapis.com)\n\n1. Go to one of the following Model Garden model cards, then click **Enable**."]]