-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Gemini API Key verification fails in Models settings despite valid API key #2892
Comments
+1 curl -H "X-Goog-Api-Key: {{YOUR_API_KEY}}"
"https://generativelanguage.googleapis.com/v1beta/models" {
"models": [
{
"name": "models/chat-bison-001",
"version": "001",
"displayName": "PaLM 2 Chat (Legacy)",
"description": "A legacy text-only model optimized for chat conversations",
"inputTokenLimit": 4096,
"outputTokenLimit": 1024,
"supportedGenerationMethods": [
"generateMessage",
"countMessageTokens"
],
"temperature": 0.25,
"topP": 0.95,
"topK": 40
},
{
"name": "models/text-bison-001",
"version": "001",
"displayName": "PaLM 2 (Legacy)",
"description": "A legacy model that understands text and generates text as an output",
"inputTokenLimit": 8196,
"outputTokenLimit": 1024,
"supportedGenerationMethods": [
"generateText",
"countTextTokens",
"createTunedTextModel"
],
"temperature": 0.7,
"topP": 0.95,
"topK": 40
},
{
"name": "models/embedding-gecko-001",
"version": "001",
"displayName": "Embedding Gecko",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 1024,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedText",
"countTextTokens"
]
},
{
"name": "models/gemini-1.0-pro-vision-latest",
"version": "001",
"displayName": "Gemini 1.0 Pro Vision",
"description": "The original Gemini 1.0 Pro Vision model version which was optimized for image understanding. Gemini 1.0 Pro Vision was deprecated on July 12, 2024. Move to a newer Gemini version.",
"inputTokenLimit": 12288,
"outputTokenLimit": 4096,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 0.4,
"topP": 1,
"topK": 32
},
{
"name": "models/gemini-pro-vision",
"version": "001",
"displayName": "Gemini 1.0 Pro Vision",
"description": "The original Gemini 1.0 Pro Vision model version which was optimized for image understanding. Gemini 1.0 Pro Vision was deprecated on July 12, 2024. Move to a newer Gemini version.",
"inputTokenLimit": 12288,
"outputTokenLimit": 4096,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 0.4,
"topP": 1,
"topK": 32
},
{
"name": "models/gemini-1.5-pro-latest",
"version": "001",
"displayName": "Gemini 1.5 Pro Latest",
"description": "Alias that points to the most recent production (non-experimental) release of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-pro-001",
"version": "001",
"displayName": "Gemini 1.5 Pro 001",
"description": "Stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens, released in May of 2024.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-pro-002",
"version": "002",
"displayName": "Gemini 1.5 Pro 002",
"description": "Stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens, released in September of 2024.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-pro",
"version": "001",
"displayName": "Gemini 1.5 Pro",
"description": "Stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens, released in May of 2024.",
"inputTokenLimit": 2000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-latest",
"version": "001",
"displayName": "Gemini 1.5 Flash Latest",
"description": "Alias that points to the most recent production (non-experimental) release of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-001",
"version": "001",
"displayName": "Gemini 1.5 Flash 001",
"description": "Stable version of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks, released in May of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-001-tuning",
"version": "001",
"displayName": "Gemini 1.5 Flash 001 Tuning",
"description": "Version of Gemini 1.5 Flash that supports tuning, our fast and versatile multimodal model for scaling across diverse tasks, released in May of 2024.",
"inputTokenLimit": 16384,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createTunedModel"
],
"temperature": 1,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash",
"version": "001",
"displayName": "Gemini 1.5 Flash",
"description": "Alias that points to the most recent stable version of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-002",
"version": "002",
"displayName": "Gemini 1.5 Flash 002",
"description": "Stable version of Gemini 1.5 Flash, our fast and versatile multimodal model for scaling across diverse tasks, released in September of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"createCachedContent"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-8b",
"version": "001",
"displayName": "Gemini 1.5 Flash-8B",
"description": "Stable version of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model, released in October of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"createCachedContent",
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-8b-001",
"version": "001",
"displayName": "Gemini 1.5 Flash-8B 001",
"description": "Stable version of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model, released in October of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"createCachedContent",
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-8b-latest",
"version": "001",
"displayName": "Gemini 1.5 Flash-8B Latest",
"description": "Alias that points to the most recent production (non-experimental) release of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model, released in October of 2024.",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"createCachedContent",
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-8b-exp-0827",
"version": "001",
"displayName": "Gemini 1.5 Flash 8B Experimental 0827",
"description": "Experimental release (August 27th, 2024) of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model. Replaced by Gemini-1.5-flash-8b-001 (stable).",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-1.5-flash-8b-exp-0924",
"version": "001",
"displayName": "Gemini 1.5 Flash 8B Experimental 0924",
"description": "Experimental release (September 24th, 2024) of Gemini 1.5 Flash-8B, our smallest and most cost effective Flash model. Replaced by Gemini-1.5-flash-8b-001 (stable).",
"inputTokenLimit": 1000000,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-exp",
"version": "2.0",
"displayName": "Gemini 2.0 Flash Experimental",
"description": "Gemini 2.0 Flash Experimental",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"bidiGenerateContent"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash",
"version": "2.0",
"displayName": "Gemini 2.0 Flash",
"description": "Gemini 2.0 Flash",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-001",
"version": "2.0",
"displayName": "Gemini 2.0 Flash 001",
"description": "Stable version of Gemini 2.0 Flash, our fast and versatile multimodal model for scaling across diverse tasks, released in January of 2025.",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-exp-image-generation",
"version": "2.0",
"displayName": "Gemini 2.0 Flash (Image Generation) Experimental",
"description": "Gemini 2.0 Flash (Image Generation) Experimental",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens",
"bidiGenerateContent"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-lite-001",
"version": "2.0",
"displayName": "Gemini 2.0 Flash-Lite 001",
"description": "Stable version of Gemini 2.0 Flash Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-lite",
"version": "2.0",
"displayName": "Gemini 2.0 Flash-Lite",
"description": "Gemini 2.0 Flash-Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-lite-preview-02-05",
"version": "preview-02-05",
"displayName": "Gemini 2.0 Flash-Lite Preview 02-05",
"description": "Preview release (February 5th, 2025) of Gemini 2.0 Flash Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-lite-preview",
"version": "preview-02-05",
"displayName": "Gemini 2.0 Flash-Lite Preview",
"description": "Preview release (February 5th, 2025) of Gemini 2.0 Flash Lite",
"inputTokenLimit": 1048576,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 40,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-pro-exp",
"version": "2.0",
"displayName": "Gemini 2.0 Pro Experimental",
"description": "Experimental release (February 5th, 2025) of Gemini 2.0 Pro",
"inputTokenLimit": 2097152,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-pro-exp-02-05",
"version": "2.0",
"displayName": "Gemini 2.0 Pro Experimental 02-05",
"description": "Experimental release (February 5th, 2025) of Gemini 2.0 Pro",
"inputTokenLimit": 2097152,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-exp-1206",
"version": "2.0",
"displayName": "Gemini Experimental 1206",
"description": "Experimental release (February 5th, 2025) of Gemini 2.0 Pro",
"inputTokenLimit": 2097152,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-thinking-exp-01-21",
"version": "2.0-exp-01-21",
"displayName": "Gemini 2.0 Flash Thinking Experimental 01-21",
"description": "Experimental release (January 21st, 2025) of Gemini 2.0 Flash Thinking",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 0.7,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-thinking-exp",
"version": "2.0-exp-01-21",
"displayName": "Gemini 2.0 Flash Thinking Experimental 01-21",
"description": "Experimental release (January 21st, 2025) of Gemini 2.0 Flash Thinking",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 0.7,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemini-2.0-flash-thinking-exp-1219",
"version": "2.0",
"displayName": "Gemini 2.0 Flash Thinking Experimental",
"description": "Gemini 2.0 Flash Thinking Experimental",
"inputTokenLimit": 1048576,
"outputTokenLimit": 65536,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 0.7,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/learnlm-1.5-pro-experimental",
"version": "001",
"displayName": "LearnLM 1.5 Pro Experimental",
"description": "Alias that points to the most recent stable version of Gemini 1.5 Pro, our mid-size multimodal model that supports up to 2 million tokens.",
"inputTokenLimit": 32767,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 64,
"maxTemperature": 2
},
{
"name": "models/gemma-3-27b-it",
"version": "001",
"displayName": "Gemma 3 27B",
"inputTokenLimit": 131072,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"generateContent",
"countTokens"
],
"temperature": 1,
"topP": 0.95,
"topK": 64
},
{
"name": "models/embedding-001",
"version": "001",
"displayName": "Embedding 001",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 2048,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent"
]
},
{
"name": "models/text-embedding-004",
"version": "004",
"displayName": "Text Embedding 004",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 2048,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent"
]
},
{
"name": "models/gemini-embedding-exp-03-07",
"version": "exp-03-07",
"displayName": "Gemini Embedding Experimental 03-07",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 8192,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent"
]
},
{
"name": "models/gemini-embedding-exp",
"version": "exp-03-07",
"displayName": "Gemini Embedding Experimental",
"description": "Obtain a distributed representation of a text.",
"inputTokenLimit": 8192,
"outputTokenLimit": 1,
"supportedGenerationMethods": [
"embedContent"
]
},
{
"name": "models/aqa",
"version": "001",
"displayName": "Model that performs Attributed Question Answering.",
"description": "Model trained to return answers to questions that are grounded in provided sources, along with estimating answerable probability.",
"inputTokenLimit": 7168,
"outputTokenLimit": 1024,
"supportedGenerationMethods": [
"generateAnswer"
],
"temperature": 0.2,
"topP": 1,
"topK": 40
},
{
"name": "models/imagen-3.0-generate-002",
"version": "002",
"displayName": "Imagen 3.0 002 model",
"description": "Vertex served Imagen 3.0 002 model",
"inputTokenLimit": 480,
"outputTokenLimit": 8192,
"supportedGenerationMethods": [
"predict"
]
}
]
} |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Issue Description
The Gemini API key verification in Cursor's Models settings shows "Invalid API Key" error, even though the API key is valid and working correctly.
Steps to Reproduce
Expected Behavior
The API key should be verified successfully as it is a valid and working key.
Actual Behavior
Verification Evidence
The API key works correctly with:
Environment
Additional Notes
The API endpoint used in Cursor's verification seems to be using "models/gemini" instead of the correct model name "gemini-1.5-pro", which might be causing the verification failure.
Note: I will also post this issue on forum.cursor.com for faster response.
The text was updated successfully, but these errors were encountered: