From 4cd2fcdc9eb2dee0a28201298d2de1beef4a7029 Mon Sep 17 00:00:00 2001
From: fern-api <115122769+fern-api[bot]@users.noreply.github.com>
Date: Fri, 13 Sep 2024 14:56:30 +0000
Subject: [PATCH] Release 0.0.1-beta26
---
pyproject.toml | 2 +-
reference.md | 112 ++---
src/gooey/__init__.py | 210 ++++++--
src/gooey/client.py | 458 ++++++++++--------
src/gooey/copilot/__init__.py | 14 +
src/gooey/copilot/client.py | 70 +--
src/gooey/copilot/types/__init__.py | 14 +
.../copilot_completion_request_asr_model.py | 23 +
...pilot_completion_request_citation_style.py | 25 +
...ilot_completion_request_embedding_model.py | 18 +
...opilot_completion_request_lipsync_model.py | 5 +
...completion_request_response_format_type.py | 5 +
...ot_completion_request_translation_model.py | 5 +
...copilot_completion_request_tts_provider.py | 7 +
src/gooey/core/client_wrapper.py | 2 +-
src/gooey/types/__init__.py | 200 ++++++--
src/gooey/types/animation_models.py | 5 -
src/gooey/types/asr_output_format.py | 5 -
src/gooey/types/asr_page_request.py | 12 +-
.../types/asr_page_request_output_format.py | 5 +
....py => asr_page_request_selected_model.py} | 2 +-
.../asr_page_request_translation_model.py | 5 +
..._eval_page_request_response_format_type.py | 5 +
src/gooey/types/combine_documents_chains.py | 5 -
...e_llm_page_request_response_format_type.py | 5 +
...compare_text2img_page_request_scheduler.py | 23 +
...t2img_page_request_selected_models_item.py | 22 +
.../types/compare_upscaler_page_request.py | 4 +-
...caler_page_request_selected_models_item.py | 7 +
src/gooey/types/create_stream_request.py | 28 +-
.../types/create_stream_request_asr_model.py | 23 +
...> create_stream_request_citation_style.py} | 2 +-
... create_stream_request_embedding_model.py} | 2 +-
.../create_stream_request_lipsync_model.py | 5 +
...ate_stream_request_response_format_type.py | 5 +
...create_stream_request_translation_model.py | 5 +
.../create_stream_request_tts_provider.py | 7 +
.../deforum_sd_page_request_selected_model.py | 5 +
src/gooey/types/doc_extract_page_request.py | 8 +-
...tract_page_request_response_format_type.py | 5 +
...extract_page_request_selected_asr_model.py | 23 +
.../doc_search_page_request_citation_style.py | 25 +
...doc_search_page_request_embedding_model.py | 18 +
...earch_page_request_response_format_type.py | 5 +
src/gooey/types/doc_summary_page_request.py | 11 +-
...mmary_page_request_response_format_type.py | 5 +
...summary_page_request_selected_asr_model.py | 23 +
...oc_summary_request_response_format_type.py | 5 +
.../doc_summary_request_selected_asr_model.py | 23 +
..._inpainting_page_request_selected_model.py | 7 +
.../embeddings_page_request_selected_model.py | 18 +
.../types/face_inpainting_page_request.py | 4 +-
..._inpainting_page_request_selected_model.py | 7 +
...google_gpt_page_request_embedding_model.py | 18 +
...e_gpt_page_request_response_format_type.py | 5 +
...e_image_gen_page_request_selected_model.py | 21 +
src/gooey/types/image_segmentation_models.py | 5 -
.../types/image_segmentation_page_request.py | 4 +-
...egmentation_page_request_selected_model.py | 5 +
src/gooey/types/img2img_page_request.py | 4 +-
..._page_request_selected_controlnet_model.py | 17 +-
..._request_selected_controlnet_model_item.py | 20 +
.../img2img_page_request_selected_model.py | 21 +
src/gooey/types/inpainting_models.py | 5 -
src/gooey/types/lipsync_models.py | 5 -
src/gooey/types/lipsync_page_request.py | 4 +-
.../lipsync_page_request_selected_model.py | 5 +
.../types/lipsync_request_selected_model.py | 5 +
src/gooey/types/lipsync_tts_page_request.py | 8 +-
...lipsync_tts_page_request_selected_model.py | 5 +
.../lipsync_tts_page_request_tts_provider.py | 7 +
.../lipsync_tts_request_selected_model.py | 5 +
...py => lipsync_tts_request_tts_provider.py} | 2 +-
.../types/object_inpainting_page_request.py | 4 +-
..._inpainting_page_request_selected_model.py | 7 +
.../types/portrait_request_selected_model.py | 5 +
.../product_image_request_selected_model.py | 5 +
.../types/qr_code_generator_page_request.py | 23 +-
...est_image_prompt_controlnet_models_item.py | 20 +
...r_code_generator_page_request_scheduler.py | 23 +
..._request_selected_controlnet_model_item.py | 20 +
...e_generator_page_request_selected_model.py | 22 +
...est_image_prompt_controlnet_models_item.py | 20 +
...dulers.py => qr_code_request_scheduler.py} | 2 +-
...request_selected_controlnet_model_item.py} | 2 +-
...s.py => qr_code_request_selected_model.py} | 2 +-
...ed_qn_a_doc_page_request_citation_style.py | 25 +
...d_qn_a_doc_page_request_embedding_model.py | 18 +
...a_doc_page_request_response_format_type.py | 5 +
...lated_qn_a_page_request_embedding_model.py | 18 +
..._qn_a_page_request_response_format_type.py | 5 +
...image_request_selected_controlnet_model.py | 17 +-
..._request_selected_controlnet_model_item.py | 20 +
... => remix_image_request_selected_model.py} | 2 +-
...emove_background_request_selected_model.py | 5 +
src/gooey/types/response_format_type.py | 5 -
...mmary_page_request_response_format_type.py | 5 +
...h_locations.py => serp_search_location.py} | 2 +-
...t_gpt_page_request_response_format_type.py | 5 +
...email_page_request_response_format_type.py | 5 +
...peech_recognition_request_output_format.py | 5 +
...eech_recognition_request_selected_model.py | 23 +
...h_recognition_request_translation_model.py | 5 +
...esize_data_request_response_format_type.py | 5 +
...thesize_data_request_selected_asr_model.py | 23 +
src/gooey/types/text2audio_models.py | 5 -
...ext_to_speech_page_request_tts_provider.py | 7 +
.../types/translate_request_selected_model.py | 5 +
src/gooey/types/translation_models.py | 5 -
src/gooey/types/translation_page_request.py | 4 +-
...translation_page_request_selected_model.py | 5 +
...> upscale_request_selected_models_item.py} | 2 +-
src/gooey/types/video_bots_page_request.py | 28 +-
.../video_bots_page_request_asr_model.py | 23 +
.../video_bots_page_request_citation_style.py | 25 +
...video_bots_page_request_embedding_model.py | 18 +
.../video_bots_page_request_lipsync_model.py | 5 +
..._bots_page_request_response_format_type.py | 5 +
...deo_bots_page_request_translation_model.py | 5 +
.../video_bots_page_request_tts_provider.py | 7 +
120 files changed, 1705 insertions(+), 512 deletions(-)
create mode 100644 src/gooey/copilot/types/copilot_completion_request_asr_model.py
create mode 100644 src/gooey/copilot/types/copilot_completion_request_citation_style.py
create mode 100644 src/gooey/copilot/types/copilot_completion_request_embedding_model.py
create mode 100644 src/gooey/copilot/types/copilot_completion_request_lipsync_model.py
create mode 100644 src/gooey/copilot/types/copilot_completion_request_response_format_type.py
create mode 100644 src/gooey/copilot/types/copilot_completion_request_translation_model.py
create mode 100644 src/gooey/copilot/types/copilot_completion_request_tts_provider.py
delete mode 100644 src/gooey/types/animation_models.py
delete mode 100644 src/gooey/types/asr_output_format.py
create mode 100644 src/gooey/types/asr_page_request_output_format.py
rename src/gooey/types/{asr_models.py => asr_page_request_selected_model.py} (91%)
create mode 100644 src/gooey/types/asr_page_request_translation_model.py
create mode 100644 src/gooey/types/bulk_eval_page_request_response_format_type.py
delete mode 100644 src/gooey/types/combine_documents_chains.py
create mode 100644 src/gooey/types/compare_llm_page_request_response_format_type.py
create mode 100644 src/gooey/types/compare_text2img_page_request_scheduler.py
create mode 100644 src/gooey/types/compare_text2img_page_request_selected_models_item.py
create mode 100644 src/gooey/types/compare_upscaler_page_request_selected_models_item.py
create mode 100644 src/gooey/types/create_stream_request_asr_model.py
rename src/gooey/types/{citation_styles.py => create_stream_request_citation_style.py} (90%)
rename src/gooey/types/{embedding_models.py => create_stream_request_embedding_model.py} (87%)
create mode 100644 src/gooey/types/create_stream_request_lipsync_model.py
create mode 100644 src/gooey/types/create_stream_request_response_format_type.py
create mode 100644 src/gooey/types/create_stream_request_translation_model.py
create mode 100644 src/gooey/types/create_stream_request_tts_provider.py
create mode 100644 src/gooey/types/deforum_sd_page_request_selected_model.py
create mode 100644 src/gooey/types/doc_extract_page_request_response_format_type.py
create mode 100644 src/gooey/types/doc_extract_page_request_selected_asr_model.py
create mode 100644 src/gooey/types/doc_search_page_request_citation_style.py
create mode 100644 src/gooey/types/doc_search_page_request_embedding_model.py
create mode 100644 src/gooey/types/doc_search_page_request_response_format_type.py
create mode 100644 src/gooey/types/doc_summary_page_request_response_format_type.py
create mode 100644 src/gooey/types/doc_summary_page_request_selected_asr_model.py
create mode 100644 src/gooey/types/doc_summary_request_response_format_type.py
create mode 100644 src/gooey/types/doc_summary_request_selected_asr_model.py
create mode 100644 src/gooey/types/email_face_inpainting_page_request_selected_model.py
create mode 100644 src/gooey/types/embeddings_page_request_selected_model.py
create mode 100644 src/gooey/types/face_inpainting_page_request_selected_model.py
create mode 100644 src/gooey/types/google_gpt_page_request_embedding_model.py
create mode 100644 src/gooey/types/google_gpt_page_request_response_format_type.py
create mode 100644 src/gooey/types/google_image_gen_page_request_selected_model.py
delete mode 100644 src/gooey/types/image_segmentation_models.py
create mode 100644 src/gooey/types/image_segmentation_page_request_selected_model.py
create mode 100644 src/gooey/types/img2img_page_request_selected_controlnet_model_item.py
create mode 100644 src/gooey/types/img2img_page_request_selected_model.py
delete mode 100644 src/gooey/types/inpainting_models.py
delete mode 100644 src/gooey/types/lipsync_models.py
create mode 100644 src/gooey/types/lipsync_page_request_selected_model.py
create mode 100644 src/gooey/types/lipsync_request_selected_model.py
create mode 100644 src/gooey/types/lipsync_tts_page_request_selected_model.py
create mode 100644 src/gooey/types/lipsync_tts_page_request_tts_provider.py
create mode 100644 src/gooey/types/lipsync_tts_request_selected_model.py
rename src/gooey/types/{text_to_speech_providers.py => lipsync_tts_request_tts_provider.py} (80%)
create mode 100644 src/gooey/types/object_inpainting_page_request_selected_model.py
create mode 100644 src/gooey/types/portrait_request_selected_model.py
create mode 100644 src/gooey/types/product_image_request_selected_model.py
create mode 100644 src/gooey/types/qr_code_generator_page_request_image_prompt_controlnet_models_item.py
create mode 100644 src/gooey/types/qr_code_generator_page_request_scheduler.py
create mode 100644 src/gooey/types/qr_code_generator_page_request_selected_controlnet_model_item.py
create mode 100644 src/gooey/types/qr_code_generator_page_request_selected_model.py
create mode 100644 src/gooey/types/qr_code_request_image_prompt_controlnet_models_item.py
rename src/gooey/types/{schedulers.py => qr_code_request_scheduler.py} (91%)
rename src/gooey/types/{control_net_models.py => qr_code_request_selected_controlnet_model_item.py} (89%)
rename src/gooey/types/{text_to_image_models.py => qr_code_request_selected_model.py} (90%)
create mode 100644 src/gooey/types/related_qn_a_doc_page_request_citation_style.py
create mode 100644 src/gooey/types/related_qn_a_doc_page_request_embedding_model.py
create mode 100644 src/gooey/types/related_qn_a_doc_page_request_response_format_type.py
create mode 100644 src/gooey/types/related_qn_a_page_request_embedding_model.py
create mode 100644 src/gooey/types/related_qn_a_page_request_response_format_type.py
create mode 100644 src/gooey/types/remix_image_request_selected_controlnet_model_item.py
rename src/gooey/types/{image_to_image_models.py => remix_image_request_selected_model.py} (89%)
create mode 100644 src/gooey/types/remove_background_request_selected_model.py
delete mode 100644 src/gooey/types/response_format_type.py
create mode 100644 src/gooey/types/seo_summary_page_request_response_format_type.py
rename src/gooey/types/{serp_search_locations.py => serp_search_location.py} (98%)
create mode 100644 src/gooey/types/smart_gpt_page_request_response_format_type.py
create mode 100644 src/gooey/types/social_lookup_email_page_request_response_format_type.py
create mode 100644 src/gooey/types/speech_recognition_request_output_format.py
create mode 100644 src/gooey/types/speech_recognition_request_selected_model.py
create mode 100644 src/gooey/types/speech_recognition_request_translation_model.py
create mode 100644 src/gooey/types/synthesize_data_request_response_format_type.py
create mode 100644 src/gooey/types/synthesize_data_request_selected_asr_model.py
delete mode 100644 src/gooey/types/text2audio_models.py
create mode 100644 src/gooey/types/text_to_speech_page_request_tts_provider.py
create mode 100644 src/gooey/types/translate_request_selected_model.py
delete mode 100644 src/gooey/types/translation_models.py
create mode 100644 src/gooey/types/translation_page_request_selected_model.py
rename src/gooey/types/{upscaler_models.py => upscale_request_selected_models_item.py} (78%)
create mode 100644 src/gooey/types/video_bots_page_request_asr_model.py
create mode 100644 src/gooey/types/video_bots_page_request_citation_style.py
create mode 100644 src/gooey/types/video_bots_page_request_embedding_model.py
create mode 100644 src/gooey/types/video_bots_page_request_lipsync_model.py
create mode 100644 src/gooey/types/video_bots_page_request_response_format_type.py
create mode 100644 src/gooey/types/video_bots_page_request_translation_model.py
create mode 100644 src/gooey/types/video_bots_page_request_tts_provider.py
diff --git a/pyproject.toml b/pyproject.toml
index 318af9f..d66b377 100644
--- a/pyproject.toml
+++ b/pyproject.toml
@@ -1,6 +1,6 @@
[tool.poetry]
name = "gooeyai"
-version = "0.0.1-beta25"
+version = "0.0.1-beta26"
description = ""
readme = "README.md"
authors = []
diff --git a/reference.md b/reference.md
index ad2abad..a067ff0 100644
--- a/reference.md
+++ b/reference.md
@@ -80,7 +80,7 @@ client.animate(
-
-**selected_model:** `typing.Optional[AnimationModels]`
+**selected_model:** `typing.Optional[DeforumSdPageRequestSelectedModel]`
@@ -308,7 +308,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**image_prompt_controlnet_models:** `typing.Optional[typing.List[ControlNetModels]]`
+**image_prompt_controlnet_models:** `typing.Optional[typing.List[QrCodeRequestImagePromptControlnetModelsItem]]`
@@ -348,7 +348,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[TextToImageModels]`
+**selected_model:** `typing.Optional[QrCodeRequestSelectedModel]`
@@ -356,7 +356,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**selected_controlnet_model:** `typing.Optional[typing.List[ControlNetModels]]`
+**selected_controlnet_model:** `typing.Optional[typing.List[QrCodeRequestSelectedControlnetModelItem]]`
@@ -412,7 +412,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**scheduler:** `typing.Optional[Schedulers]`
+**scheduler:** `typing.Optional[QrCodeRequestScheduler]`
@@ -605,7 +605,7 @@ client.seo_people_also_ask(
-
-**embedding_model:** `typing.Optional[EmbeddingModels]`
+**embedding_model:** `typing.Optional[RelatedQnAPageRequestEmbeddingModel]`
@@ -666,7 +666,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[RelatedQnAPageRequestResponseFormatType]`
@@ -674,7 +674,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**serp_search_location:** `typing.Optional[SerpSearchLocations]`
+**serp_search_location:** `typing.Optional[SerpSearchLocation]`
@@ -893,7 +893,7 @@ client.seo_content(
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[SeoSummaryPageRequestResponseFormatType]`
@@ -901,7 +901,7 @@ client.seo_content(
-
-**serp_search_location:** `typing.Optional[SerpSearchLocations]`
+**serp_search_location:** `typing.Optional[SerpSearchLocation]`
@@ -1086,7 +1086,7 @@ client.web_search_llm(
-
-**embedding_model:** `typing.Optional[EmbeddingModels]`
+**embedding_model:** `typing.Optional[GoogleGptPageRequestEmbeddingModel]`
@@ -1147,7 +1147,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[GoogleGptPageRequestResponseFormatType]`
@@ -1155,7 +1155,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**serp_search_location:** `typing.Optional[SerpSearchLocations]`
+**serp_search_location:** `typing.Optional[SerpSearchLocation]`
@@ -1331,7 +1331,7 @@ client.personalize_email(
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[SocialLookupEmailPageRequestResponseFormatType]`
@@ -1647,7 +1647,7 @@ Aggregate using one or more operations. Uses [pandas](https://pandas.pydata.org/
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[BulkEvalPageRequestResponseFormatType]`
@@ -1753,7 +1753,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**selected_asr_model:** `typing.Optional[AsrModels]`
+**selected_asr_model:** `typing.Optional[SynthesizeDataRequestSelectedAsrModel]`
@@ -1835,7 +1835,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[SynthesizeDataRequestResponseFormatType]`
@@ -1977,7 +1977,7 @@ client.llm()
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[CompareLlmPageRequestResponseFormatType]`
@@ -2121,7 +2121,7 @@ client.rag(
-
-**embedding_model:** `typing.Optional[EmbeddingModels]`
+**embedding_model:** `typing.Optional[DocSearchPageRequestEmbeddingModel]`
@@ -2166,7 +2166,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**citation_style:** `typing.Optional[CitationStyles]`
+**citation_style:** `typing.Optional[DocSearchPageRequestCitationStyle]`
@@ -2214,7 +2214,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[DocSearchPageRequestResponseFormatType]`
@@ -2382,7 +2382,7 @@ client.smart_gpt(
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[SmartGptPageRequestResponseFormatType]`
@@ -2502,7 +2502,7 @@ typing.List[core.File]` — See core.File for more documentation
-
-**chain_type:** `typing.Optional[CombineDocumentsChains]`
+**chain_type:** `typing.Optional[typing.Literal["map_reduce"]]`
@@ -2510,7 +2510,7 @@ typing.List[core.File]` — See core.File for more documentation
-
-**selected_asr_model:** `typing.Optional[AsrModels]`
+**selected_asr_model:** `typing.Optional[DocSummaryRequestSelectedAsrModel]`
@@ -2566,7 +2566,7 @@ typing.List[core.File]` — See core.File for more documentation
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[DocSummaryRequestResponseFormatType]`
@@ -2780,7 +2780,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[LipsyncModels]`
+**selected_model:** `typing.Optional[LipsyncRequestSelectedModel]`
@@ -2886,7 +2886,7 @@ client.lipsync_tts(
-
-**tts_provider:** `typing.Optional[TextToSpeechProviders]`
+**tts_provider:** `typing.Optional[LipsyncTtsRequestTtsProvider]`
@@ -3080,7 +3080,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[LipsyncModels]`
+**selected_model:** `typing.Optional[LipsyncTtsRequestSelectedModel]`
@@ -3176,7 +3176,7 @@ client.text_to_speech(
-
-**tts_provider:** `typing.Optional[TextToSpeechProviders]`
+**tts_provider:** `typing.Optional[TextToSpeechPageRequestTtsProvider]`
@@ -3408,7 +3408,7 @@ typing.List[core.File]` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[AsrModels]`
+**selected_model:** `typing.Optional[SpeechRecognitionRequestSelectedModel]`
@@ -3424,7 +3424,7 @@ typing.List[core.File]` — See core.File for more documentation
-
-**translation_model:** `typing.Optional[TranslationModels]`
+**translation_model:** `typing.Optional[SpeechRecognitionRequestTranslationModel]`
@@ -3432,7 +3432,7 @@ typing.List[core.File]` — See core.File for more documentation
-
-**output_format:** `typing.Optional[AsrOutputFormat]`
+**output_format:** `typing.Optional[SpeechRecognitionRequestOutputFormat]`
@@ -3618,7 +3618,7 @@ client.text_to_music(
-
-**selected_models:** `typing.Optional[typing.Sequence[Text2AudioModels]]`
+**selected_models:** `typing.Optional[typing.Sequence[typing.Literal["audio_ldm"]]]`
@@ -3712,7 +3712,7 @@ client.translate()
-
-**selected_model:** `typing.Optional[TranslationModels]`
+**selected_model:** `typing.Optional[TranslateRequestSelectedModel]`
@@ -3842,7 +3842,7 @@ core.File` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[ImageToImageModels]`
+**selected_model:** `typing.Optional[RemixImageRequestSelectedModel]`
@@ -4106,7 +4106,7 @@ client.text_to_image(
-
-**selected_models:** `typing.Optional[typing.Sequence[TextToImageModels]]`
+**selected_models:** `typing.Optional[typing.Sequence[CompareText2ImgPageRequestSelectedModelsItem]]`
@@ -4114,7 +4114,7 @@ client.text_to_image(
-
-**scheduler:** `typing.Optional[Schedulers]`
+**scheduler:** `typing.Optional[CompareText2ImgPageRequestScheduler]`
@@ -4268,7 +4268,7 @@ core.File` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[InpaintingModels]`
+**selected_model:** `typing.Optional[ProductImageRequestSelectedModel]`
@@ -4462,7 +4462,7 @@ core.File` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[InpaintingModels]`
+**selected_model:** `typing.Optional[PortraitRequestSelectedModel]`
@@ -4663,7 +4663,7 @@ client.image_from_email(
-
-**selected_model:** `typing.Optional[InpaintingModels]`
+**selected_model:** `typing.Optional[EmailFaceInpaintingPageRequestSelectedModel]`
@@ -4896,7 +4896,7 @@ client.image_from_web_search(
-
-**serp_search_location:** `typing.Optional[SerpSearchLocations]`
+**serp_search_location:** `typing.Optional[SerpSearchLocation]`
@@ -4912,7 +4912,7 @@ client.image_from_web_search(
-
-**selected_model:** `typing.Optional[ImageToImageModels]`
+**selected_model:** `typing.Optional[GoogleImageGenPageRequestSelectedModel]`
@@ -5072,7 +5072,7 @@ core.File` — See core.File for more documentation
-
-**selected_model:** `typing.Optional[ImageSegmentationModels]`
+**selected_model:** `typing.Optional[RemoveBackgroundRequestSelectedModel]`
@@ -5236,7 +5236,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**selected_models:** `typing.Optional[typing.List[UpscalerModels]]`
+**selected_models:** `typing.Optional[typing.List[UpscaleRequestSelectedModelsItem]]`
@@ -5340,7 +5340,7 @@ client.embed(
-
-**selected_model:** `typing.Optional[EmbeddingModels]`
+**selected_model:** `typing.Optional[EmbeddingsPageRequestSelectedModel]`
@@ -5484,7 +5484,7 @@ client.seo_people_also_ask_doc(
-
-**embedding_model:** `typing.Optional[EmbeddingModels]`
+**embedding_model:** `typing.Optional[RelatedQnADocPageRequestEmbeddingModel]`
@@ -5529,7 +5529,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**citation_style:** `typing.Optional[CitationStyles]`
+**citation_style:** `typing.Optional[RelatedQnADocPageRequestCitationStyle]`
@@ -5577,7 +5577,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[RelatedQnADocPageRequestResponseFormatType]`
@@ -5585,7 +5585,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**serp_search_location:** `typing.Optional[SerpSearchLocations]`
+**serp_search_location:** `typing.Optional[SerpSearchLocation]`
@@ -5876,7 +5876,7 @@ typing.Optional[typing.List[core.File]]` — See core.File for more documentatio
-
-**embedding_model:** `typing.Optional[EmbeddingModels]`
+**embedding_model:** `typing.Optional[CopilotCompletionRequestEmbeddingModel]`
@@ -5897,7 +5897,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**citation_style:** `typing.Optional[CitationStyles]`
+**citation_style:** `typing.Optional[CopilotCompletionRequestCitationStyle]`
@@ -5913,7 +5913,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**asr_model:** `typing.Optional[AsrModels]` — Choose a model to transcribe incoming audio messages to text.
+**asr_model:** `typing.Optional[CopilotCompletionRequestAsrModel]` — Choose a model to transcribe incoming audio messages to text.
@@ -5929,7 +5929,7 @@ Generally speaking, dense embeddings excel at understanding the context of the q
-
-**translation_model:** `typing.Optional[TranslationModels]`
+**translation_model:** `typing.Optional[CopilotCompletionRequestTranslationModel]`
@@ -5965,7 +5965,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**lipsync_model:** `typing.Optional[LipsyncModels]`
+**lipsync_model:** `typing.Optional[CopilotCompletionRequestLipsyncModel]`
@@ -6021,7 +6021,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**response_format_type:** `typing.Optional[ResponseFormatType]`
+**response_format_type:** `typing.Optional[CopilotCompletionRequestResponseFormatType]`
@@ -6029,7 +6029,7 @@ typing.Optional[core.File]` — See core.File for more documentation
-
-**tts_provider:** `typing.Optional[TextToSpeechProviders]`
+**tts_provider:** `typing.Optional[CopilotCompletionRequestTtsProvider]`
diff --git a/src/gooey/__init__.py b/src/gooey/__init__.py
index b9e8980..86305e9 100644
--- a/src/gooey/__init__.py
+++ b/src/gooey/__init__.py
@@ -5,21 +5,22 @@
AggFunctionFunction,
AggFunctionResult,
AggFunctionResultFunction,
- AnimationModels,
AnimationPrompt,
AsrChunk,
- AsrModels,
- AsrOutputFormat,
AsrOutputJson,
AsrPageOutput,
AsrPageOutputOutputTextItem,
AsrPageRequest,
+ AsrPageRequestOutputFormat,
+ AsrPageRequestSelectedModel,
+ AsrPageRequestTranslationModel,
AsrPageStatusResponse,
AsyncApiResponseModelV3,
BalanceResponse,
BotBroadcastFilters,
BotBroadcastRequestModel,
BulkEvalPageOutput,
+ BulkEvalPageRequestResponseFormatType,
BulkEvalPageStatusResponse,
BulkRunnerPageOutput,
BulkRunnerPageRequest,
@@ -32,18 +33,19 @@
ChyronPlantPageOutput,
ChyronPlantPageRequest,
ChyronPlantPageStatusResponse,
- CitationStyles,
- CombineDocumentsChains,
CompareLlmPageOutput,
+ CompareLlmPageRequestResponseFormatType,
CompareLlmPageStatusResponse,
CompareText2ImgPageOutput,
+ CompareText2ImgPageRequestScheduler,
+ CompareText2ImgPageRequestSelectedModelsItem,
CompareText2ImgPageStatusResponse,
CompareUpscalerPageOutput,
CompareUpscalerPageRequest,
+ CompareUpscalerPageRequestSelectedModelsItem,
CompareUpscalerPageStatusResponse,
ConsoleLogs,
ConsoleLogsLevel,
- ControlNetModels,
ConversationEntry,
ConversationEntryContent,
ConversationEntryContentItem,
@@ -52,28 +54,47 @@
ConversationEntryRole,
ConversationStart,
CreateStreamRequest,
+ CreateStreamRequestAsrModel,
+ CreateStreamRequestCitationStyle,
+ CreateStreamRequestEmbeddingModel,
+ CreateStreamRequestLipsyncModel,
CreateStreamRequestOpenaiTtsModel,
CreateStreamRequestOpenaiVoiceName,
+ CreateStreamRequestResponseFormatType,
+ CreateStreamRequestTranslationModel,
+ CreateStreamRequestTtsProvider,
CreateStreamResponse,
DeforumSdPageOutput,
+ DeforumSdPageRequestSelectedModel,
DeforumSdPageStatusResponse,
DocExtractPageOutput,
DocExtractPageRequest,
+ DocExtractPageRequestResponseFormatType,
+ DocExtractPageRequestSelectedAsrModel,
DocExtractPageStatusResponse,
DocSearchPageOutput,
+ DocSearchPageRequestCitationStyle,
+ DocSearchPageRequestEmbeddingModel,
DocSearchPageRequestKeywordQuery,
+ DocSearchPageRequestResponseFormatType,
DocSearchPageStatusResponse,
DocSummaryPageOutput,
DocSummaryPageRequest,
+ DocSummaryPageRequestResponseFormatType,
+ DocSummaryPageRequestSelectedAsrModel,
DocSummaryPageStatusResponse,
+ DocSummaryRequestResponseFormatType,
+ DocSummaryRequestSelectedAsrModel,
EmailFaceInpaintingPageOutput,
+ EmailFaceInpaintingPageRequestSelectedModel,
EmailFaceInpaintingPageStatusResponse,
- EmbeddingModels,
EmbeddingsPageOutput,
+ EmbeddingsPageRequestSelectedModel,
EmbeddingsPageStatusResponse,
EvalPrompt,
FaceInpaintingPageOutput,
FaceInpaintingPageRequest,
+ FaceInpaintingPageRequestSelectedModel,
FaceInpaintingPageStatusResponse,
FinalResponse,
FunctionsPageOutput,
@@ -81,60 +102,86 @@
GenericErrorResponse,
GenericErrorResponseDetail,
GoogleGptPageOutput,
+ GoogleGptPageRequestEmbeddingModel,
+ GoogleGptPageRequestResponseFormatType,
GoogleGptPageStatusResponse,
GoogleImageGenPageOutput,
+ GoogleImageGenPageRequestSelectedModel,
GoogleImageGenPageStatusResponse,
HttpValidationError,
- ImageSegmentationModels,
ImageSegmentationPageOutput,
ImageSegmentationPageRequest,
+ ImageSegmentationPageRequestSelectedModel,
ImageSegmentationPageStatusResponse,
- ImageToImageModels,
ImageUrl,
ImageUrlDetail,
Img2ImgPageOutput,
Img2ImgPageRequest,
Img2ImgPageRequestSelectedControlnetModel,
+ Img2ImgPageRequestSelectedControlnetModelItem,
+ Img2ImgPageRequestSelectedModel,
Img2ImgPageStatusResponse,
- InpaintingModels,
LargeLanguageModels,
LetterWriterPageOutput,
LetterWriterPageRequest,
LetterWriterPageStatusResponse,
- LipsyncModels,
LipsyncPageOutput,
LipsyncPageRequest,
+ LipsyncPageRequestSelectedModel,
LipsyncPageStatusResponse,
+ LipsyncRequestSelectedModel,
LipsyncTtsPageOutput,
LipsyncTtsPageRequest,
LipsyncTtsPageRequestOpenaiTtsModel,
LipsyncTtsPageRequestOpenaiVoiceName,
+ LipsyncTtsPageRequestSelectedModel,
+ LipsyncTtsPageRequestTtsProvider,
LipsyncTtsPageStatusResponse,
LipsyncTtsRequestOpenaiTtsModel,
LipsyncTtsRequestOpenaiVoiceName,
+ LipsyncTtsRequestSelectedModel,
+ LipsyncTtsRequestTtsProvider,
LlmTools,
MessagePart,
ObjectInpaintingPageOutput,
ObjectInpaintingPageRequest,
+ ObjectInpaintingPageRequestSelectedModel,
ObjectInpaintingPageStatusResponse,
+ PortraitRequestSelectedModel,
+ ProductImageRequestSelectedModel,
PromptTreeNode,
PromptTreeNodePrompt,
QrCodeGeneratorPageOutput,
QrCodeGeneratorPageRequest,
+ QrCodeGeneratorPageRequestImagePromptControlnetModelsItem,
+ QrCodeGeneratorPageRequestScheduler,
+ QrCodeGeneratorPageRequestSelectedControlnetModelItem,
+ QrCodeGeneratorPageRequestSelectedModel,
QrCodeGeneratorPageStatusResponse,
+ QrCodeRequestImagePromptControlnetModelsItem,
+ QrCodeRequestScheduler,
+ QrCodeRequestSelectedControlnetModelItem,
+ QrCodeRequestSelectedModel,
RecipeFunction,
RecipeFunctionTrigger,
RecipeRunState,
RelatedDocSearchResponse,
RelatedGoogleGptResponse,
RelatedQnADocPageOutput,
+ RelatedQnADocPageRequestCitationStyle,
+ RelatedQnADocPageRequestEmbeddingModel,
RelatedQnADocPageRequestKeywordQuery,
+ RelatedQnADocPageRequestResponseFormatType,
RelatedQnADocPageStatusResponse,
RelatedQnAPageOutput,
+ RelatedQnAPageRequestEmbeddingModel,
+ RelatedQnAPageRequestResponseFormatType,
RelatedQnAPageStatusResponse,
RemixImageRequestSelectedControlnetModel,
+ RemixImageRequestSelectedControlnetModelItem,
+ RemixImageRequestSelectedModel,
+ RemoveBackgroundRequestSelectedModel,
ReplyButton,
- ResponseFormatType,
ResponseModel,
ResponseModelFinalKeywordQuery,
ResponseModelFinalPrompt,
@@ -143,32 +190,38 @@
RunStart,
SadTalkerSettings,
SadTalkerSettingsPreprocess,
- Schedulers,
SearchReference,
SeoSummaryPageOutput,
+ SeoSummaryPageRequestResponseFormatType,
SeoSummaryPageStatusResponse,
- SerpSearchLocations,
+ SerpSearchLocation,
SerpSearchType,
SmartGptPageOutput,
+ SmartGptPageRequestResponseFormatType,
SmartGptPageStatusResponse,
SocialLookupEmailPageOutput,
+ SocialLookupEmailPageRequestResponseFormatType,
SocialLookupEmailPageStatusResponse,
+ SpeechRecognitionRequestOutputFormat,
+ SpeechRecognitionRequestSelectedModel,
+ SpeechRecognitionRequestTranslationModel,
StreamError,
- Text2AudioModels,
+ SynthesizeDataRequestResponseFormatType,
+ SynthesizeDataRequestSelectedAsrModel,
Text2AudioPageOutput,
Text2AudioPageStatusResponse,
- TextToImageModels,
TextToSpeechPageOutput,
TextToSpeechPageRequestOpenaiTtsModel,
TextToSpeechPageRequestOpenaiVoiceName,
+ TextToSpeechPageRequestTtsProvider,
TextToSpeechPageStatusResponse,
- TextToSpeechProviders,
TrainingDataModel,
- TranslationModels,
+ TranslateRequestSelectedModel,
TranslationPageOutput,
TranslationPageRequest,
+ TranslationPageRequestSelectedModel,
TranslationPageStatusResponse,
- UpscalerModels,
+ UpscaleRequestSelectedModelsItem,
ValidationError,
ValidationErrorLocItem,
Vcard,
@@ -176,24 +229,38 @@
VideoBotsPageOutputFinalKeywordQuery,
VideoBotsPageOutputFinalPrompt,
VideoBotsPageRequest,
+ VideoBotsPageRequestAsrModel,
+ VideoBotsPageRequestCitationStyle,
+ VideoBotsPageRequestEmbeddingModel,
VideoBotsPageRequestFunctionsItem,
VideoBotsPageRequestFunctionsItemTrigger,
+ VideoBotsPageRequestLipsyncModel,
VideoBotsPageRequestOpenaiTtsModel,
VideoBotsPageRequestOpenaiVoiceName,
+ VideoBotsPageRequestResponseFormatType,
VideoBotsPageRequestSadtalkerSettings,
VideoBotsPageRequestSadtalkerSettingsPreprocess,
+ VideoBotsPageRequestTranslationModel,
+ VideoBotsPageRequestTtsProvider,
VideoBotsPageStatusResponse,
)
from .errors import PaymentRequiredError, TooManyRequestsError, UnprocessableEntityError
from . import copilot
from .client import AsyncGooey, Gooey
from .copilot import (
+ CopilotCompletionRequestAsrModel,
+ CopilotCompletionRequestCitationStyle,
+ CopilotCompletionRequestEmbeddingModel,
CopilotCompletionRequestFunctionsItem,
CopilotCompletionRequestFunctionsItemTrigger,
+ CopilotCompletionRequestLipsyncModel,
CopilotCompletionRequestOpenaiTtsModel,
CopilotCompletionRequestOpenaiVoiceName,
+ CopilotCompletionRequestResponseFormatType,
CopilotCompletionRequestSadtalkerSettings,
CopilotCompletionRequestSadtalkerSettingsPreprocess,
+ CopilotCompletionRequestTranslationModel,
+ CopilotCompletionRequestTtsProvider,
)
from .environment import GooeyEnvironment
from .version import __version__
@@ -203,15 +270,15 @@
"AggFunctionFunction",
"AggFunctionResult",
"AggFunctionResultFunction",
- "AnimationModels",
"AnimationPrompt",
"AsrChunk",
- "AsrModels",
- "AsrOutputFormat",
"AsrOutputJson",
"AsrPageOutput",
"AsrPageOutputOutputTextItem",
"AsrPageRequest",
+ "AsrPageRequestOutputFormat",
+ "AsrPageRequestSelectedModel",
+ "AsrPageRequestTranslationModel",
"AsrPageStatusResponse",
"AsyncApiResponseModelV3",
"AsyncGooey",
@@ -219,6 +286,7 @@
"BotBroadcastFilters",
"BotBroadcastRequestModel",
"BulkEvalPageOutput",
+ "BulkEvalPageRequestResponseFormatType",
"BulkEvalPageStatusResponse",
"BulkRunnerPageOutput",
"BulkRunnerPageRequest",
@@ -231,18 +299,19 @@
"ChyronPlantPageOutput",
"ChyronPlantPageRequest",
"ChyronPlantPageStatusResponse",
- "CitationStyles",
- "CombineDocumentsChains",
"CompareLlmPageOutput",
+ "CompareLlmPageRequestResponseFormatType",
"CompareLlmPageStatusResponse",
"CompareText2ImgPageOutput",
+ "CompareText2ImgPageRequestScheduler",
+ "CompareText2ImgPageRequestSelectedModelsItem",
"CompareText2ImgPageStatusResponse",
"CompareUpscalerPageOutput",
"CompareUpscalerPageRequest",
+ "CompareUpscalerPageRequestSelectedModelsItem",
"CompareUpscalerPageStatusResponse",
"ConsoleLogs",
"ConsoleLogsLevel",
- "ControlNetModels",
"ConversationEntry",
"ConversationEntryContent",
"ConversationEntryContentItem",
@@ -250,35 +319,61 @@
"ConversationEntryContentItem_Text",
"ConversationEntryRole",
"ConversationStart",
+ "CopilotCompletionRequestAsrModel",
+ "CopilotCompletionRequestCitationStyle",
+ "CopilotCompletionRequestEmbeddingModel",
"CopilotCompletionRequestFunctionsItem",
"CopilotCompletionRequestFunctionsItemTrigger",
+ "CopilotCompletionRequestLipsyncModel",
"CopilotCompletionRequestOpenaiTtsModel",
"CopilotCompletionRequestOpenaiVoiceName",
+ "CopilotCompletionRequestResponseFormatType",
"CopilotCompletionRequestSadtalkerSettings",
"CopilotCompletionRequestSadtalkerSettingsPreprocess",
+ "CopilotCompletionRequestTranslationModel",
+ "CopilotCompletionRequestTtsProvider",
"CreateStreamRequest",
+ "CreateStreamRequestAsrModel",
+ "CreateStreamRequestCitationStyle",
+ "CreateStreamRequestEmbeddingModel",
+ "CreateStreamRequestLipsyncModel",
"CreateStreamRequestOpenaiTtsModel",
"CreateStreamRequestOpenaiVoiceName",
+ "CreateStreamRequestResponseFormatType",
+ "CreateStreamRequestTranslationModel",
+ "CreateStreamRequestTtsProvider",
"CreateStreamResponse",
"DeforumSdPageOutput",
+ "DeforumSdPageRequestSelectedModel",
"DeforumSdPageStatusResponse",
"DocExtractPageOutput",
"DocExtractPageRequest",
+ "DocExtractPageRequestResponseFormatType",
+ "DocExtractPageRequestSelectedAsrModel",
"DocExtractPageStatusResponse",
"DocSearchPageOutput",
+ "DocSearchPageRequestCitationStyle",
+ "DocSearchPageRequestEmbeddingModel",
"DocSearchPageRequestKeywordQuery",
+ "DocSearchPageRequestResponseFormatType",
"DocSearchPageStatusResponse",
"DocSummaryPageOutput",
"DocSummaryPageRequest",
+ "DocSummaryPageRequestResponseFormatType",
+ "DocSummaryPageRequestSelectedAsrModel",
"DocSummaryPageStatusResponse",
+ "DocSummaryRequestResponseFormatType",
+ "DocSummaryRequestSelectedAsrModel",
"EmailFaceInpaintingPageOutput",
+ "EmailFaceInpaintingPageRequestSelectedModel",
"EmailFaceInpaintingPageStatusResponse",
- "EmbeddingModels",
"EmbeddingsPageOutput",
+ "EmbeddingsPageRequestSelectedModel",
"EmbeddingsPageStatusResponse",
"EvalPrompt",
"FaceInpaintingPageOutput",
"FaceInpaintingPageRequest",
+ "FaceInpaintingPageRequestSelectedModel",
"FaceInpaintingPageStatusResponse",
"FinalResponse",
"FunctionsPageOutput",
@@ -288,61 +383,87 @@
"Gooey",
"GooeyEnvironment",
"GoogleGptPageOutput",
+ "GoogleGptPageRequestEmbeddingModel",
+ "GoogleGptPageRequestResponseFormatType",
"GoogleGptPageStatusResponse",
"GoogleImageGenPageOutput",
+ "GoogleImageGenPageRequestSelectedModel",
"GoogleImageGenPageStatusResponse",
"HttpValidationError",
- "ImageSegmentationModels",
"ImageSegmentationPageOutput",
"ImageSegmentationPageRequest",
+ "ImageSegmentationPageRequestSelectedModel",
"ImageSegmentationPageStatusResponse",
- "ImageToImageModels",
"ImageUrl",
"ImageUrlDetail",
"Img2ImgPageOutput",
"Img2ImgPageRequest",
"Img2ImgPageRequestSelectedControlnetModel",
+ "Img2ImgPageRequestSelectedControlnetModelItem",
+ "Img2ImgPageRequestSelectedModel",
"Img2ImgPageStatusResponse",
- "InpaintingModels",
"LargeLanguageModels",
"LetterWriterPageOutput",
"LetterWriterPageRequest",
"LetterWriterPageStatusResponse",
- "LipsyncModels",
"LipsyncPageOutput",
"LipsyncPageRequest",
+ "LipsyncPageRequestSelectedModel",
"LipsyncPageStatusResponse",
+ "LipsyncRequestSelectedModel",
"LipsyncTtsPageOutput",
"LipsyncTtsPageRequest",
"LipsyncTtsPageRequestOpenaiTtsModel",
"LipsyncTtsPageRequestOpenaiVoiceName",
+ "LipsyncTtsPageRequestSelectedModel",
+ "LipsyncTtsPageRequestTtsProvider",
"LipsyncTtsPageStatusResponse",
"LipsyncTtsRequestOpenaiTtsModel",
"LipsyncTtsRequestOpenaiVoiceName",
+ "LipsyncTtsRequestSelectedModel",
+ "LipsyncTtsRequestTtsProvider",
"LlmTools",
"MessagePart",
"ObjectInpaintingPageOutput",
"ObjectInpaintingPageRequest",
+ "ObjectInpaintingPageRequestSelectedModel",
"ObjectInpaintingPageStatusResponse",
"PaymentRequiredError",
+ "PortraitRequestSelectedModel",
+ "ProductImageRequestSelectedModel",
"PromptTreeNode",
"PromptTreeNodePrompt",
"QrCodeGeneratorPageOutput",
"QrCodeGeneratorPageRequest",
+ "QrCodeGeneratorPageRequestImagePromptControlnetModelsItem",
+ "QrCodeGeneratorPageRequestScheduler",
+ "QrCodeGeneratorPageRequestSelectedControlnetModelItem",
+ "QrCodeGeneratorPageRequestSelectedModel",
"QrCodeGeneratorPageStatusResponse",
+ "QrCodeRequestImagePromptControlnetModelsItem",
+ "QrCodeRequestScheduler",
+ "QrCodeRequestSelectedControlnetModelItem",
+ "QrCodeRequestSelectedModel",
"RecipeFunction",
"RecipeFunctionTrigger",
"RecipeRunState",
"RelatedDocSearchResponse",
"RelatedGoogleGptResponse",
"RelatedQnADocPageOutput",
+ "RelatedQnADocPageRequestCitationStyle",
+ "RelatedQnADocPageRequestEmbeddingModel",
"RelatedQnADocPageRequestKeywordQuery",
+ "RelatedQnADocPageRequestResponseFormatType",
"RelatedQnADocPageStatusResponse",
"RelatedQnAPageOutput",
+ "RelatedQnAPageRequestEmbeddingModel",
+ "RelatedQnAPageRequestResponseFormatType",
"RelatedQnAPageStatusResponse",
"RemixImageRequestSelectedControlnetModel",
+ "RemixImageRequestSelectedControlnetModelItem",
+ "RemixImageRequestSelectedModel",
+ "RemoveBackgroundRequestSelectedModel",
"ReplyButton",
- "ResponseFormatType",
"ResponseModel",
"ResponseModelFinalKeywordQuery",
"ResponseModelFinalPrompt",
@@ -351,34 +472,40 @@
"RunStart",
"SadTalkerSettings",
"SadTalkerSettingsPreprocess",
- "Schedulers",
"SearchReference",
"SeoSummaryPageOutput",
+ "SeoSummaryPageRequestResponseFormatType",
"SeoSummaryPageStatusResponse",
- "SerpSearchLocations",
+ "SerpSearchLocation",
"SerpSearchType",
"SmartGptPageOutput",
+ "SmartGptPageRequestResponseFormatType",
"SmartGptPageStatusResponse",
"SocialLookupEmailPageOutput",
+ "SocialLookupEmailPageRequestResponseFormatType",
"SocialLookupEmailPageStatusResponse",
+ "SpeechRecognitionRequestOutputFormat",
+ "SpeechRecognitionRequestSelectedModel",
+ "SpeechRecognitionRequestTranslationModel",
"StreamError",
- "Text2AudioModels",
+ "SynthesizeDataRequestResponseFormatType",
+ "SynthesizeDataRequestSelectedAsrModel",
"Text2AudioPageOutput",
"Text2AudioPageStatusResponse",
- "TextToImageModels",
"TextToSpeechPageOutput",
"TextToSpeechPageRequestOpenaiTtsModel",
"TextToSpeechPageRequestOpenaiVoiceName",
+ "TextToSpeechPageRequestTtsProvider",
"TextToSpeechPageStatusResponse",
- "TextToSpeechProviders",
"TooManyRequestsError",
"TrainingDataModel",
- "TranslationModels",
+ "TranslateRequestSelectedModel",
"TranslationPageOutput",
"TranslationPageRequest",
+ "TranslationPageRequestSelectedModel",
"TranslationPageStatusResponse",
"UnprocessableEntityError",
- "UpscalerModels",
+ "UpscaleRequestSelectedModelsItem",
"ValidationError",
"ValidationErrorLocItem",
"Vcard",
@@ -386,12 +513,19 @@
"VideoBotsPageOutputFinalKeywordQuery",
"VideoBotsPageOutputFinalPrompt",
"VideoBotsPageRequest",
+ "VideoBotsPageRequestAsrModel",
+ "VideoBotsPageRequestCitationStyle",
+ "VideoBotsPageRequestEmbeddingModel",
"VideoBotsPageRequestFunctionsItem",
"VideoBotsPageRequestFunctionsItemTrigger",
+ "VideoBotsPageRequestLipsyncModel",
"VideoBotsPageRequestOpenaiTtsModel",
"VideoBotsPageRequestOpenaiVoiceName",
+ "VideoBotsPageRequestResponseFormatType",
"VideoBotsPageRequestSadtalkerSettings",
"VideoBotsPageRequestSadtalkerSettingsPreprocess",
+ "VideoBotsPageRequestTranslationModel",
+ "VideoBotsPageRequestTtsProvider",
"VideoBotsPageStatusResponse",
"__version__",
"copilot",
diff --git a/src/gooey/client.py b/src/gooey/client.py
index 10c5a84..6767f27 100644
--- a/src/gooey/client.py
+++ b/src/gooey/client.py
@@ -9,7 +9,7 @@
from .copilot.client import CopilotClient
from .types.animation_prompt import AnimationPrompt
from .types.recipe_function import RecipeFunction
-from .types.animation_models import AnimationModels
+from .types.deforum_sd_page_request_selected_model import DeforumSdPageRequestSelectedModel
from .types.run_settings import RunSettings
from .core.request_options import RequestOptions
from .types.deforum_sd_page_output import DeforumSdPageOutput
@@ -22,64 +22,88 @@
from json.decoder import JSONDecodeError
from . import core
from .types.vcard import Vcard
-from .types.control_net_models import ControlNetModels
-from .types.text_to_image_models import TextToImageModels
-from .types.schedulers import Schedulers
+from .types.qr_code_request_image_prompt_controlnet_models_item import QrCodeRequestImagePromptControlnetModelsItem
+from .types.qr_code_request_selected_model import QrCodeRequestSelectedModel
+from .types.qr_code_request_selected_controlnet_model_item import QrCodeRequestSelectedControlnetModelItem
+from .types.qr_code_request_scheduler import QrCodeRequestScheduler
from .types.qr_code_generator_page_output import QrCodeGeneratorPageOutput
from .types.large_language_models import LargeLanguageModels
-from .types.embedding_models import EmbeddingModels
-from .types.response_format_type import ResponseFormatType
-from .types.serp_search_locations import SerpSearchLocations
+from .types.related_qn_a_page_request_embedding_model import RelatedQnAPageRequestEmbeddingModel
+from .types.related_qn_a_page_request_response_format_type import RelatedQnAPageRequestResponseFormatType
+from .types.serp_search_location import SerpSearchLocation
from .types.serp_search_type import SerpSearchType
from .types.related_qn_a_page_output import RelatedQnAPageOutput
+from .types.seo_summary_page_request_response_format_type import SeoSummaryPageRequestResponseFormatType
from .types.seo_summary_page_output import SeoSummaryPageOutput
+from .types.google_gpt_page_request_embedding_model import GoogleGptPageRequestEmbeddingModel
+from .types.google_gpt_page_request_response_format_type import GoogleGptPageRequestResponseFormatType
from .types.google_gpt_page_output import GoogleGptPageOutput
+from .types.social_lookup_email_page_request_response_format_type import SocialLookupEmailPageRequestResponseFormatType
from .types.social_lookup_email_page_output import SocialLookupEmailPageOutput
from .types.bulk_runner_page_output import BulkRunnerPageOutput
from .types.eval_prompt import EvalPrompt
from .types.agg_function import AggFunction
+from .types.bulk_eval_page_request_response_format_type import BulkEvalPageRequestResponseFormatType
from .types.bulk_eval_page_output import BulkEvalPageOutput
-from .types.asr_models import AsrModels
+from .types.synthesize_data_request_selected_asr_model import SynthesizeDataRequestSelectedAsrModel
+from .types.synthesize_data_request_response_format_type import SynthesizeDataRequestResponseFormatType
from .types.doc_extract_page_output import DocExtractPageOutput
+from .types.compare_llm_page_request_response_format_type import CompareLlmPageRequestResponseFormatType
from .types.compare_llm_page_output import CompareLlmPageOutput
from .types.doc_search_page_request_keyword_query import DocSearchPageRequestKeywordQuery
-from .types.citation_styles import CitationStyles
+from .types.doc_search_page_request_embedding_model import DocSearchPageRequestEmbeddingModel
+from .types.doc_search_page_request_citation_style import DocSearchPageRequestCitationStyle
+from .types.doc_search_page_request_response_format_type import DocSearchPageRequestResponseFormatType
from .types.doc_search_page_output import DocSearchPageOutput
+from .types.smart_gpt_page_request_response_format_type import SmartGptPageRequestResponseFormatType
from .types.smart_gpt_page_output import SmartGptPageOutput
-from .types.combine_documents_chains import CombineDocumentsChains
+from .types.doc_summary_request_selected_asr_model import DocSummaryRequestSelectedAsrModel
+from .types.doc_summary_request_response_format_type import DocSummaryRequestResponseFormatType
from .types.doc_summary_page_output import DocSummaryPageOutput
from .types.functions_page_output import FunctionsPageOutput
from .types.sad_talker_settings import SadTalkerSettings
-from .types.lipsync_models import LipsyncModels
+from .types.lipsync_request_selected_model import LipsyncRequestSelectedModel
from .types.lipsync_page_output import LipsyncPageOutput
-from .types.text_to_speech_providers import TextToSpeechProviders
+from .types.lipsync_tts_request_tts_provider import LipsyncTtsRequestTtsProvider
from .types.lipsync_tts_request_openai_voice_name import LipsyncTtsRequestOpenaiVoiceName
from .types.lipsync_tts_request_openai_tts_model import LipsyncTtsRequestOpenaiTtsModel
+from .types.lipsync_tts_request_selected_model import LipsyncTtsRequestSelectedModel
from .types.lipsync_tts_page_output import LipsyncTtsPageOutput
+from .types.text_to_speech_page_request_tts_provider import TextToSpeechPageRequestTtsProvider
from .types.text_to_speech_page_request_openai_voice_name import TextToSpeechPageRequestOpenaiVoiceName
from .types.text_to_speech_page_request_openai_tts_model import TextToSpeechPageRequestOpenaiTtsModel
from .types.text_to_speech_page_output import TextToSpeechPageOutput
-from .types.translation_models import TranslationModels
-from .types.asr_output_format import AsrOutputFormat
+from .types.speech_recognition_request_selected_model import SpeechRecognitionRequestSelectedModel
+from .types.speech_recognition_request_translation_model import SpeechRecognitionRequestTranslationModel
+from .types.speech_recognition_request_output_format import SpeechRecognitionRequestOutputFormat
from .types.asr_page_output import AsrPageOutput
-from .types.text2audio_models import Text2AudioModels
from .types.text2audio_page_output import Text2AudioPageOutput
+from .types.translate_request_selected_model import TranslateRequestSelectedModel
from .types.translation_page_output import TranslationPageOutput
-from .types.image_to_image_models import ImageToImageModels
+from .types.remix_image_request_selected_model import RemixImageRequestSelectedModel
from .types.remix_image_request_selected_controlnet_model import RemixImageRequestSelectedControlnetModel
from .types.img2img_page_output import Img2ImgPageOutput
+from .types.compare_text2img_page_request_selected_models_item import CompareText2ImgPageRequestSelectedModelsItem
+from .types.compare_text2img_page_request_scheduler import CompareText2ImgPageRequestScheduler
from .types.compare_text2img_page_output import CompareText2ImgPageOutput
-from .types.inpainting_models import InpaintingModels
+from .types.product_image_request_selected_model import ProductImageRequestSelectedModel
from .types.object_inpainting_page_output import ObjectInpaintingPageOutput
+from .types.portrait_request_selected_model import PortraitRequestSelectedModel
from .types.face_inpainting_page_output import FaceInpaintingPageOutput
+from .types.email_face_inpainting_page_request_selected_model import EmailFaceInpaintingPageRequestSelectedModel
from .types.email_face_inpainting_page_output import EmailFaceInpaintingPageOutput
+from .types.google_image_gen_page_request_selected_model import GoogleImageGenPageRequestSelectedModel
from .types.google_image_gen_page_output import GoogleImageGenPageOutput
-from .types.image_segmentation_models import ImageSegmentationModels
+from .types.remove_background_request_selected_model import RemoveBackgroundRequestSelectedModel
from .types.image_segmentation_page_output import ImageSegmentationPageOutput
-from .types.upscaler_models import UpscalerModels
+from .types.upscale_request_selected_models_item import UpscaleRequestSelectedModelsItem
from .types.compare_upscaler_page_output import CompareUpscalerPageOutput
+from .types.embeddings_page_request_selected_model import EmbeddingsPageRequestSelectedModel
from .types.embeddings_page_output import EmbeddingsPageOutput
from .types.related_qn_a_doc_page_request_keyword_query import RelatedQnADocPageRequestKeywordQuery
+from .types.related_qn_a_doc_page_request_embedding_model import RelatedQnADocPageRequestEmbeddingModel
+from .types.related_qn_a_doc_page_request_citation_style import RelatedQnADocPageRequestCitationStyle
+from .types.related_qn_a_doc_page_request_response_format_type import RelatedQnADocPageRequestResponseFormatType
from .types.related_qn_a_doc_page_output import RelatedQnADocPageOutput
from .types.balance_response import BalanceResponse
from .core.client_wrapper import AsyncClientWrapper
@@ -159,7 +183,7 @@ def animate(
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
max_frames: typing.Optional[int] = OMIT,
- selected_model: typing.Optional[AnimationModels] = OMIT,
+ selected_model: typing.Optional[DeforumSdPageRequestSelectedModel] = OMIT,
animation_mode: typing.Optional[str] = OMIT,
zoom: typing.Optional[str] = OMIT,
translation_x: typing.Optional[str] = OMIT,
@@ -186,7 +210,7 @@ def animate(
max_frames : typing.Optional[int]
- selected_model : typing.Optional[AnimationModels]
+ selected_model : typing.Optional[DeforumSdPageRequestSelectedModel]
animation_mode : typing.Optional[str]
@@ -316,20 +340,22 @@ def qr_code(
use_url_shortener: typing.Optional[bool] = None,
negative_prompt: typing.Optional[str] = None,
image_prompt: typing.Optional[str] = None,
- image_prompt_controlnet_models: typing.Optional[typing.List[ControlNetModels]] = None,
+ image_prompt_controlnet_models: typing.Optional[
+ typing.List[QrCodeRequestImagePromptControlnetModelsItem]
+ ] = None,
image_prompt_strength: typing.Optional[float] = None,
image_prompt_scale: typing.Optional[float] = None,
image_prompt_pos_x: typing.Optional[float] = None,
image_prompt_pos_y: typing.Optional[float] = None,
- selected_model: typing.Optional[TextToImageModels] = None,
- selected_controlnet_model: typing.Optional[typing.List[ControlNetModels]] = None,
+ selected_model: typing.Optional[QrCodeRequestSelectedModel] = None,
+ selected_controlnet_model: typing.Optional[typing.List[QrCodeRequestSelectedControlnetModelItem]] = None,
output_width: typing.Optional[int] = None,
output_height: typing.Optional[int] = None,
guidance_scale: typing.Optional[float] = None,
controlnet_conditioning_scale: typing.Optional[typing.List[float]] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[int] = None,
- scheduler: typing.Optional[Schedulers] = None,
+ scheduler: typing.Optional[QrCodeRequestScheduler] = None,
seed: typing.Optional[int] = None,
obj_scale: typing.Optional[float] = None,
obj_pos_x: typing.Optional[float] = None,
@@ -365,7 +391,7 @@ def qr_code(
image_prompt : typing.Optional[str]
- image_prompt_controlnet_models : typing.Optional[typing.List[ControlNetModels]]
+ image_prompt_controlnet_models : typing.Optional[typing.List[QrCodeRequestImagePromptControlnetModelsItem]]
image_prompt_strength : typing.Optional[float]
@@ -375,9 +401,9 @@ def qr_code(
image_prompt_pos_y : typing.Optional[float]
- selected_model : typing.Optional[TextToImageModels]
+ selected_model : typing.Optional[QrCodeRequestSelectedModel]
- selected_controlnet_model : typing.Optional[typing.List[ControlNetModels]]
+ selected_controlnet_model : typing.Optional[typing.List[QrCodeRequestSelectedControlnetModelItem]]
output_width : typing.Optional[int]
@@ -391,7 +417,7 @@ def qr_code(
quality : typing.Optional[int]
- scheduler : typing.Optional[Schedulers]
+ scheduler : typing.Optional[QrCodeRequestScheduler]
seed : typing.Optional[int]
@@ -523,15 +549,15 @@ def seo_people_also_ask(
max_references: typing.Optional[int] = OMIT,
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[RelatedQnAPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[RelatedQnAPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -566,7 +592,7 @@ def seo_people_also_ask(
scroll_jump : typing.Optional[int]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[RelatedQnAPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -584,9 +610,9 @@ def seo_people_also_ask(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[RelatedQnAPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
@@ -716,8 +742,8 @@ def seo_content(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[SeoSummaryPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -759,9 +785,9 @@ def seo_content(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SeoSummaryPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
@@ -886,15 +912,15 @@ def web_search_llm(
max_references: typing.Optional[int] = OMIT,
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[GoogleGptPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[GoogleGptPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -929,7 +955,7 @@ def web_search_llm(
scroll_jump : typing.Optional[int]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[GoogleGptPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -947,9 +973,9 @@ def web_search_llm(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[GoogleGptPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
@@ -1074,7 +1100,7 @@ def personalize_email(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[SocialLookupEmailPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> SocialLookupEmailPageOutput:
@@ -1104,7 +1130,7 @@ def personalize_email(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SocialLookupEmailPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -1345,7 +1371,7 @@ def eval(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[BulkEvalPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> BulkEvalPageOutput:
@@ -1389,7 +1415,7 @@ def eval(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[BulkEvalPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -1488,7 +1514,7 @@ def synthesize_data(
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
sheet_url: typing.Optional[core.File] = None,
- selected_asr_model: typing.Optional[AsrModels] = None,
+ selected_asr_model: typing.Optional[SynthesizeDataRequestSelectedAsrModel] = None,
google_translate_target: typing.Optional[str] = None,
glossary_document: typing.Optional[core.File] = None,
task_instructions: typing.Optional[str] = None,
@@ -1498,7 +1524,7 @@ def synthesize_data(
quality: typing.Optional[float] = None,
max_tokens: typing.Optional[int] = None,
sampling_temperature: typing.Optional[float] = None,
- response_format_type: typing.Optional[ResponseFormatType] = None,
+ response_format_type: typing.Optional[SynthesizeDataRequestResponseFormatType] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
) -> DocExtractPageOutput:
@@ -1518,7 +1544,7 @@ def synthesize_data(
sheet_url : typing.Optional[core.File]
See core.File for more documentation
- selected_asr_model : typing.Optional[AsrModels]
+ selected_asr_model : typing.Optional[SynthesizeDataRequestSelectedAsrModel]
google_translate_target : typing.Optional[str]
@@ -1539,7 +1565,7 @@ def synthesize_data(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SynthesizeDataRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -1646,7 +1672,7 @@ def llm(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[CompareLlmPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> CompareLlmPageOutput:
@@ -1674,7 +1700,7 @@ def llm(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[CompareLlmPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -1774,18 +1800,18 @@ def rag(
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
doc_extract_url: typing.Optional[str] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[DocSearchPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
task_instructions: typing.Optional[str] = OMIT,
query_instructions: typing.Optional[str] = OMIT,
selected_model: typing.Optional[LargeLanguageModels] = OMIT,
- citation_style: typing.Optional[CitationStyles] = OMIT,
+ citation_style: typing.Optional[DocSearchPageRequestCitationStyle] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[DocSearchPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> DocSearchPageOutput:
@@ -1813,7 +1839,7 @@ def rag(
doc_extract_url : typing.Optional[str]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[DocSearchPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -1827,7 +1853,7 @@ def rag(
selected_model : typing.Optional[LargeLanguageModels]
- citation_style : typing.Optional[CitationStyles]
+ citation_style : typing.Optional[DocSearchPageRequestCitationStyle]
avoid_repetition : typing.Optional[bool]
@@ -1839,7 +1865,7 @@ def rag(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[DocSearchPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -1955,7 +1981,7 @@ def smart_gpt(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[SmartGptPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> SmartGptPageOutput:
@@ -1989,7 +2015,7 @@ def smart_gpt(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SmartGptPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -2091,15 +2117,15 @@ def doc_summary(
task_instructions: typing.Optional[str] = None,
merge_instructions: typing.Optional[str] = None,
selected_model: typing.Optional[LargeLanguageModels] = None,
- chain_type: typing.Optional[CombineDocumentsChains] = None,
- selected_asr_model: typing.Optional[AsrModels] = None,
+ chain_type: typing.Optional[typing.Literal["map_reduce"]] = None,
+ selected_asr_model: typing.Optional[DocSummaryRequestSelectedAsrModel] = None,
google_translate_target: typing.Optional[str] = None,
avoid_repetition: typing.Optional[bool] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[float] = None,
max_tokens: typing.Optional[int] = None,
sampling_temperature: typing.Optional[float] = None,
- response_format_type: typing.Optional[ResponseFormatType] = None,
+ response_format_type: typing.Optional[DocSummaryRequestResponseFormatType] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
) -> DocSummaryPageOutput:
@@ -2122,9 +2148,9 @@ def doc_summary(
selected_model : typing.Optional[LargeLanguageModels]
- chain_type : typing.Optional[CombineDocumentsChains]
+ chain_type : typing.Optional[typing.Literal["map_reduce"]]
- selected_asr_model : typing.Optional[AsrModels]
+ selected_asr_model : typing.Optional[DocSummaryRequestSelectedAsrModel]
google_translate_target : typing.Optional[str]
@@ -2138,7 +2164,7 @@ def doc_summary(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[DocSummaryRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -2341,7 +2367,7 @@ def lipsync(
face_padding_left: typing.Optional[int] = None,
face_padding_right: typing.Optional[int] = None,
sadtalker_settings: typing.Optional[SadTalkerSettings] = None,
- selected_model: typing.Optional[LipsyncModels] = None,
+ selected_model: typing.Optional[LipsyncRequestSelectedModel] = None,
input_audio: typing.Optional[core.File] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
@@ -2369,7 +2395,7 @@ def lipsync(
sadtalker_settings : typing.Optional[SadTalkerSettings]
- selected_model : typing.Optional[LipsyncModels]
+ selected_model : typing.Optional[LipsyncRequestSelectedModel]
input_audio : typing.Optional[core.File]
See core.File for more documentation
@@ -2468,7 +2494,7 @@ def lipsync_tts(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
- tts_provider: typing.Optional[TextToSpeechProviders] = None,
+ tts_provider: typing.Optional[LipsyncTtsRequestTtsProvider] = None,
uberduck_voice_name: typing.Optional[str] = None,
uberduck_speaking_rate: typing.Optional[float] = None,
google_voice_name: typing.Optional[str] = None,
@@ -2492,7 +2518,7 @@ def lipsync_tts(
face_padding_left: typing.Optional[int] = None,
face_padding_right: typing.Optional[int] = None,
sadtalker_settings: typing.Optional[SadTalkerSettings] = None,
- selected_model: typing.Optional[LipsyncModels] = None,
+ selected_model: typing.Optional[LipsyncTtsRequestSelectedModel] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
) -> LipsyncTtsPageOutput:
@@ -2508,7 +2534,7 @@ def lipsync_tts(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- tts_provider : typing.Optional[TextToSpeechProviders]
+ tts_provider : typing.Optional[LipsyncTtsRequestTtsProvider]
uberduck_voice_name : typing.Optional[str]
@@ -2558,7 +2584,7 @@ def lipsync_tts(
sadtalker_settings : typing.Optional[SadTalkerSettings]
- selected_model : typing.Optional[LipsyncModels]
+ selected_model : typing.Optional[LipsyncTtsRequestSelectedModel]
settings : typing.Optional[RunSettings]
@@ -2674,7 +2700,7 @@ def text_to_speech(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
- tts_provider: typing.Optional[TextToSpeechProviders] = OMIT,
+ tts_provider: typing.Optional[TextToSpeechPageRequestTtsProvider] = OMIT,
uberduck_voice_name: typing.Optional[str] = OMIT,
uberduck_speaking_rate: typing.Optional[float] = OMIT,
google_voice_name: typing.Optional[str] = OMIT,
@@ -2707,7 +2733,7 @@ def text_to_speech(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- tts_provider : typing.Optional[TextToSpeechProviders]
+ tts_provider : typing.Optional[TextToSpeechPageRequestTtsProvider]
uberduck_voice_name : typing.Optional[str]
@@ -2849,10 +2875,10 @@ def speech_recognition(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
- selected_model: typing.Optional[AsrModels] = None,
+ selected_model: typing.Optional[SpeechRecognitionRequestSelectedModel] = None,
language: typing.Optional[str] = None,
- translation_model: typing.Optional[TranslationModels] = None,
- output_format: typing.Optional[AsrOutputFormat] = None,
+ translation_model: typing.Optional[SpeechRecognitionRequestTranslationModel] = None,
+ output_format: typing.Optional[SpeechRecognitionRequestOutputFormat] = None,
google_translate_target: typing.Optional[str] = None,
translation_source: typing.Optional[str] = None,
translation_target: typing.Optional[str] = None,
@@ -2873,13 +2899,13 @@ def speech_recognition(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- selected_model : typing.Optional[AsrModels]
+ selected_model : typing.Optional[SpeechRecognitionRequestSelectedModel]
language : typing.Optional[str]
- translation_model : typing.Optional[TranslationModels]
+ translation_model : typing.Optional[SpeechRecognitionRequestTranslationModel]
- output_format : typing.Optional[AsrOutputFormat]
+ output_format : typing.Optional[SpeechRecognitionRequestOutputFormat]
google_translate_target : typing.Optional[str]
use `translation_model` & `translation_target` instead.
@@ -2993,7 +3019,7 @@ def text_to_music(
guidance_scale: typing.Optional[float] = OMIT,
seed: typing.Optional[int] = OMIT,
sd2upscaling: typing.Optional[bool] = OMIT,
- selected_models: typing.Optional[typing.Sequence[Text2AudioModels]] = OMIT,
+ selected_models: typing.Optional[typing.Sequence[typing.Literal["audio_ldm"]]] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> Text2AudioPageOutput:
@@ -3023,7 +3049,7 @@ def text_to_music(
sd2upscaling : typing.Optional[bool]
- selected_models : typing.Optional[typing.Sequence[Text2AudioModels]]
+ selected_models : typing.Optional[typing.Sequence[typing.Literal["audio_ldm"]]]
settings : typing.Optional[RunSettings]
@@ -3120,7 +3146,7 @@ def translate(
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
texts: typing.Optional[typing.List[str]] = None,
- selected_model: typing.Optional[TranslationModels] = None,
+ selected_model: typing.Optional[TranslateRequestSelectedModel] = None,
translation_source: typing.Optional[str] = None,
translation_target: typing.Optional[str] = None,
glossary_document: typing.Optional[core.File] = None,
@@ -3139,7 +3165,7 @@ def translate(
texts : typing.Optional[typing.List[str]]
- selected_model : typing.Optional[TranslationModels]
+ selected_model : typing.Optional[TranslateRequestSelectedModel]
translation_source : typing.Optional[str]
@@ -3240,7 +3266,7 @@ def remix_image(
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
text_prompt: typing.Optional[str] = None,
- selected_model: typing.Optional[ImageToImageModels] = None,
+ selected_model: typing.Optional[RemixImageRequestSelectedModel] = None,
selected_controlnet_model: typing.Optional[RemixImageRequestSelectedControlnetModel] = None,
negative_prompt: typing.Optional[str] = None,
num_outputs: typing.Optional[int] = None,
@@ -3270,7 +3296,7 @@ def remix_image(
text_prompt : typing.Optional[str]
- selected_model : typing.Optional[ImageToImageModels]
+ selected_model : typing.Optional[RemixImageRequestSelectedModel]
selected_controlnet_model : typing.Optional[RemixImageRequestSelectedControlnetModel]
@@ -3404,8 +3430,8 @@ def text_to_image(
guidance_scale: typing.Optional[float] = OMIT,
seed: typing.Optional[int] = OMIT,
sd2upscaling: typing.Optional[bool] = OMIT,
- selected_models: typing.Optional[typing.Sequence[TextToImageModels]] = OMIT,
- scheduler: typing.Optional[Schedulers] = OMIT,
+ selected_models: typing.Optional[typing.Sequence[CompareText2ImgPageRequestSelectedModelsItem]] = OMIT,
+ scheduler: typing.Optional[CompareText2ImgPageRequestScheduler] = OMIT,
edit_instruction: typing.Optional[str] = OMIT,
image_guidance_scale: typing.Optional[float] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
@@ -3443,9 +3469,9 @@ def text_to_image(
sd2upscaling : typing.Optional[bool]
- selected_models : typing.Optional[typing.Sequence[TextToImageModels]]
+ selected_models : typing.Optional[typing.Sequence[CompareText2ImgPageRequestSelectedModelsItem]]
- scheduler : typing.Optional[Schedulers]
+ scheduler : typing.Optional[CompareText2ImgPageRequestScheduler]
edit_instruction : typing.Optional[str]
@@ -3557,7 +3583,7 @@ def product_image(
obj_pos_x: typing.Optional[float] = None,
obj_pos_y: typing.Optional[float] = None,
mask_threshold: typing.Optional[float] = None,
- selected_model: typing.Optional[InpaintingModels] = None,
+ selected_model: typing.Optional[ProductImageRequestSelectedModel] = None,
negative_prompt: typing.Optional[str] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[int] = None,
@@ -3592,7 +3618,7 @@ def product_image(
mask_threshold : typing.Optional[float]
- selected_model : typing.Optional[InpaintingModels]
+ selected_model : typing.Optional[ProductImageRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -3717,7 +3743,7 @@ def portrait(
face_scale: typing.Optional[float] = None,
face_pos_x: typing.Optional[float] = None,
face_pos_y: typing.Optional[float] = None,
- selected_model: typing.Optional[InpaintingModels] = None,
+ selected_model: typing.Optional[PortraitRequestSelectedModel] = None,
negative_prompt: typing.Optional[str] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[int] = None,
@@ -3750,7 +3776,7 @@ def portrait(
face_pos_y : typing.Optional[float]
- selected_model : typing.Optional[InpaintingModels]
+ selected_model : typing.Optional[PortraitRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -3875,7 +3901,7 @@ def image_from_email(
face_scale: typing.Optional[float] = OMIT,
face_pos_x: typing.Optional[float] = OMIT,
face_pos_y: typing.Optional[float] = OMIT,
- selected_model: typing.Optional[InpaintingModels] = OMIT,
+ selected_model: typing.Optional[EmailFaceInpaintingPageRequestSelectedModel] = OMIT,
negative_prompt: typing.Optional[str] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[int] = OMIT,
@@ -3917,7 +3943,7 @@ def image_from_email(
face_pos_y : typing.Optional[float]
- selected_model : typing.Optional[InpaintingModels]
+ selected_model : typing.Optional[EmailFaceInpaintingPageRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -4062,9 +4088,9 @@ def image_from_web_search(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
- selected_model: typing.Optional[ImageToImageModels] = OMIT,
+ selected_model: typing.Optional[GoogleImageGenPageRequestSelectedModel] = OMIT,
negative_prompt: typing.Optional[str] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[int] = OMIT,
@@ -4090,12 +4116,12 @@ def image_from_web_search(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
- selected_model : typing.Optional[ImageToImageModels]
+ selected_model : typing.Optional[GoogleImageGenPageRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -4213,7 +4239,7 @@ def remove_background(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
- selected_model: typing.Optional[ImageSegmentationModels] = None,
+ selected_model: typing.Optional[RemoveBackgroundRequestSelectedModel] = None,
mask_threshold: typing.Optional[float] = None,
rect_persepective_transform: typing.Optional[bool] = None,
reflection_opacity: typing.Optional[float] = None,
@@ -4236,7 +4262,7 @@ def remove_background(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- selected_model : typing.Optional[ImageSegmentationModels]
+ selected_model : typing.Optional[RemoveBackgroundRequestSelectedModel]
mask_threshold : typing.Optional[float]
@@ -4346,7 +4372,7 @@ def upscale(
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
input_image: typing.Optional[core.File] = None,
input_video: typing.Optional[core.File] = None,
- selected_models: typing.Optional[typing.List[UpscalerModels]] = None,
+ selected_models: typing.Optional[typing.List[UpscaleRequestSelectedModelsItem]] = None,
selected_bg_model: typing.Optional[typing.Literal["real_esrgan_x2"]] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
@@ -4370,7 +4396,7 @@ def upscale(
input_video : typing.Optional[core.File]
See core.File for more documentation
- selected_models : typing.Optional[typing.List[UpscalerModels]]
+ selected_models : typing.Optional[typing.List[UpscaleRequestSelectedModelsItem]]
selected_bg_model : typing.Optional[typing.Literal["real_esrgan_x2"]]
@@ -4467,7 +4493,7 @@ def embed(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
- selected_model: typing.Optional[EmbeddingModels] = OMIT,
+ selected_model: typing.Optional[EmbeddingsPageRequestSelectedModel] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> EmbeddingsPageOutput:
@@ -4483,7 +4509,7 @@ def embed(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- selected_model : typing.Optional[EmbeddingModels]
+ selected_model : typing.Optional[EmbeddingsPageRequestSelectedModel]
settings : typing.Optional[RunSettings]
@@ -4579,19 +4605,19 @@ def seo_people_also_ask_doc(
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
doc_extract_url: typing.Optional[str] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[RelatedQnADocPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
task_instructions: typing.Optional[str] = OMIT,
query_instructions: typing.Optional[str] = OMIT,
selected_model: typing.Optional[LargeLanguageModels] = OMIT,
- citation_style: typing.Optional[CitationStyles] = OMIT,
+ citation_style: typing.Optional[RelatedQnADocPageRequestCitationStyle] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[RelatedQnADocPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -4622,7 +4648,7 @@ def seo_people_also_ask_doc(
doc_extract_url : typing.Optional[str]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[RelatedQnADocPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -4636,7 +4662,7 @@ def seo_people_also_ask_doc(
selected_model : typing.Optional[LargeLanguageModels]
- citation_style : typing.Optional[CitationStyles]
+ citation_style : typing.Optional[RelatedQnADocPageRequestCitationStyle]
avoid_repetition : typing.Optional[bool]
@@ -4648,9 +4674,9 @@ def seo_people_also_ask_doc(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[RelatedQnADocPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
@@ -4873,7 +4899,7 @@ async def animate(
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
max_frames: typing.Optional[int] = OMIT,
- selected_model: typing.Optional[AnimationModels] = OMIT,
+ selected_model: typing.Optional[DeforumSdPageRequestSelectedModel] = OMIT,
animation_mode: typing.Optional[str] = OMIT,
zoom: typing.Optional[str] = OMIT,
translation_x: typing.Optional[str] = OMIT,
@@ -4900,7 +4926,7 @@ async def animate(
max_frames : typing.Optional[int]
- selected_model : typing.Optional[AnimationModels]
+ selected_model : typing.Optional[DeforumSdPageRequestSelectedModel]
animation_mode : typing.Optional[str]
@@ -5038,20 +5064,22 @@ async def qr_code(
use_url_shortener: typing.Optional[bool] = None,
negative_prompt: typing.Optional[str] = None,
image_prompt: typing.Optional[str] = None,
- image_prompt_controlnet_models: typing.Optional[typing.List[ControlNetModels]] = None,
+ image_prompt_controlnet_models: typing.Optional[
+ typing.List[QrCodeRequestImagePromptControlnetModelsItem]
+ ] = None,
image_prompt_strength: typing.Optional[float] = None,
image_prompt_scale: typing.Optional[float] = None,
image_prompt_pos_x: typing.Optional[float] = None,
image_prompt_pos_y: typing.Optional[float] = None,
- selected_model: typing.Optional[TextToImageModels] = None,
- selected_controlnet_model: typing.Optional[typing.List[ControlNetModels]] = None,
+ selected_model: typing.Optional[QrCodeRequestSelectedModel] = None,
+ selected_controlnet_model: typing.Optional[typing.List[QrCodeRequestSelectedControlnetModelItem]] = None,
output_width: typing.Optional[int] = None,
output_height: typing.Optional[int] = None,
guidance_scale: typing.Optional[float] = None,
controlnet_conditioning_scale: typing.Optional[typing.List[float]] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[int] = None,
- scheduler: typing.Optional[Schedulers] = None,
+ scheduler: typing.Optional[QrCodeRequestScheduler] = None,
seed: typing.Optional[int] = None,
obj_scale: typing.Optional[float] = None,
obj_pos_x: typing.Optional[float] = None,
@@ -5087,7 +5115,7 @@ async def qr_code(
image_prompt : typing.Optional[str]
- image_prompt_controlnet_models : typing.Optional[typing.List[ControlNetModels]]
+ image_prompt_controlnet_models : typing.Optional[typing.List[QrCodeRequestImagePromptControlnetModelsItem]]
image_prompt_strength : typing.Optional[float]
@@ -5097,9 +5125,9 @@ async def qr_code(
image_prompt_pos_y : typing.Optional[float]
- selected_model : typing.Optional[TextToImageModels]
+ selected_model : typing.Optional[QrCodeRequestSelectedModel]
- selected_controlnet_model : typing.Optional[typing.List[ControlNetModels]]
+ selected_controlnet_model : typing.Optional[typing.List[QrCodeRequestSelectedControlnetModelItem]]
output_width : typing.Optional[int]
@@ -5113,7 +5141,7 @@ async def qr_code(
quality : typing.Optional[int]
- scheduler : typing.Optional[Schedulers]
+ scheduler : typing.Optional[QrCodeRequestScheduler]
seed : typing.Optional[int]
@@ -5253,15 +5281,15 @@ async def seo_people_also_ask(
max_references: typing.Optional[int] = OMIT,
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[RelatedQnAPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[RelatedQnAPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -5296,7 +5324,7 @@ async def seo_people_also_ask(
scroll_jump : typing.Optional[int]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[RelatedQnAPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -5314,9 +5342,9 @@ async def seo_people_also_ask(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[RelatedQnAPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
@@ -5454,8 +5482,8 @@ async def seo_content(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[SeoSummaryPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -5497,9 +5525,9 @@ async def seo_content(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SeoSummaryPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
@@ -5632,15 +5660,15 @@ async def web_search_llm(
max_references: typing.Optional[int] = OMIT,
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[GoogleGptPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[GoogleGptPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -5675,7 +5703,7 @@ async def web_search_llm(
scroll_jump : typing.Optional[int]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[GoogleGptPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -5693,9 +5721,9 @@ async def web_search_llm(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[GoogleGptPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
@@ -5828,7 +5856,7 @@ async def personalize_email(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[SocialLookupEmailPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> SocialLookupEmailPageOutput:
@@ -5858,7 +5886,7 @@ async def personalize_email(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SocialLookupEmailPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -6115,7 +6143,7 @@ async def eval(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[BulkEvalPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> BulkEvalPageOutput:
@@ -6159,7 +6187,7 @@ async def eval(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[BulkEvalPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -6266,7 +6294,7 @@ async def synthesize_data(
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
sheet_url: typing.Optional[core.File] = None,
- selected_asr_model: typing.Optional[AsrModels] = None,
+ selected_asr_model: typing.Optional[SynthesizeDataRequestSelectedAsrModel] = None,
google_translate_target: typing.Optional[str] = None,
glossary_document: typing.Optional[core.File] = None,
task_instructions: typing.Optional[str] = None,
@@ -6276,7 +6304,7 @@ async def synthesize_data(
quality: typing.Optional[float] = None,
max_tokens: typing.Optional[int] = None,
sampling_temperature: typing.Optional[float] = None,
- response_format_type: typing.Optional[ResponseFormatType] = None,
+ response_format_type: typing.Optional[SynthesizeDataRequestResponseFormatType] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
) -> DocExtractPageOutput:
@@ -6296,7 +6324,7 @@ async def synthesize_data(
sheet_url : typing.Optional[core.File]
See core.File for more documentation
- selected_asr_model : typing.Optional[AsrModels]
+ selected_asr_model : typing.Optional[SynthesizeDataRequestSelectedAsrModel]
google_translate_target : typing.Optional[str]
@@ -6317,7 +6345,7 @@ async def synthesize_data(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SynthesizeDataRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -6432,7 +6460,7 @@ async def llm(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[CompareLlmPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> CompareLlmPageOutput:
@@ -6460,7 +6488,7 @@ async def llm(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[CompareLlmPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -6568,18 +6596,18 @@ async def rag(
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
doc_extract_url: typing.Optional[str] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[DocSearchPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
task_instructions: typing.Optional[str] = OMIT,
query_instructions: typing.Optional[str] = OMIT,
selected_model: typing.Optional[LargeLanguageModels] = OMIT,
- citation_style: typing.Optional[CitationStyles] = OMIT,
+ citation_style: typing.Optional[DocSearchPageRequestCitationStyle] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[DocSearchPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> DocSearchPageOutput:
@@ -6607,7 +6635,7 @@ async def rag(
doc_extract_url : typing.Optional[str]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[DocSearchPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -6621,7 +6649,7 @@ async def rag(
selected_model : typing.Optional[LargeLanguageModels]
- citation_style : typing.Optional[CitationStyles]
+ citation_style : typing.Optional[DocSearchPageRequestCitationStyle]
avoid_repetition : typing.Optional[bool]
@@ -6633,7 +6661,7 @@ async def rag(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[DocSearchPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -6757,7 +6785,7 @@ async def smart_gpt(
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
+ response_format_type: typing.Optional[SmartGptPageRequestResponseFormatType] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> SmartGptPageOutput:
@@ -6791,7 +6819,7 @@ async def smart_gpt(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[SmartGptPageRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -6901,15 +6929,15 @@ async def doc_summary(
task_instructions: typing.Optional[str] = None,
merge_instructions: typing.Optional[str] = None,
selected_model: typing.Optional[LargeLanguageModels] = None,
- chain_type: typing.Optional[CombineDocumentsChains] = None,
- selected_asr_model: typing.Optional[AsrModels] = None,
+ chain_type: typing.Optional[typing.Literal["map_reduce"]] = None,
+ selected_asr_model: typing.Optional[DocSummaryRequestSelectedAsrModel] = None,
google_translate_target: typing.Optional[str] = None,
avoid_repetition: typing.Optional[bool] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[float] = None,
max_tokens: typing.Optional[int] = None,
sampling_temperature: typing.Optional[float] = None,
- response_format_type: typing.Optional[ResponseFormatType] = None,
+ response_format_type: typing.Optional[DocSummaryRequestResponseFormatType] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
) -> DocSummaryPageOutput:
@@ -6932,9 +6960,9 @@ async def doc_summary(
selected_model : typing.Optional[LargeLanguageModels]
- chain_type : typing.Optional[CombineDocumentsChains]
+ chain_type : typing.Optional[typing.Literal["map_reduce"]]
- selected_asr_model : typing.Optional[AsrModels]
+ selected_asr_model : typing.Optional[DocSummaryRequestSelectedAsrModel]
google_translate_target : typing.Optional[str]
@@ -6948,7 +6976,7 @@ async def doc_summary(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[DocSummaryRequestResponseFormatType]
settings : typing.Optional[RunSettings]
@@ -7167,7 +7195,7 @@ async def lipsync(
face_padding_left: typing.Optional[int] = None,
face_padding_right: typing.Optional[int] = None,
sadtalker_settings: typing.Optional[SadTalkerSettings] = None,
- selected_model: typing.Optional[LipsyncModels] = None,
+ selected_model: typing.Optional[LipsyncRequestSelectedModel] = None,
input_audio: typing.Optional[core.File] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
@@ -7195,7 +7223,7 @@ async def lipsync(
sadtalker_settings : typing.Optional[SadTalkerSettings]
- selected_model : typing.Optional[LipsyncModels]
+ selected_model : typing.Optional[LipsyncRequestSelectedModel]
input_audio : typing.Optional[core.File]
See core.File for more documentation
@@ -7302,7 +7330,7 @@ async def lipsync_tts(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
- tts_provider: typing.Optional[TextToSpeechProviders] = None,
+ tts_provider: typing.Optional[LipsyncTtsRequestTtsProvider] = None,
uberduck_voice_name: typing.Optional[str] = None,
uberduck_speaking_rate: typing.Optional[float] = None,
google_voice_name: typing.Optional[str] = None,
@@ -7326,7 +7354,7 @@ async def lipsync_tts(
face_padding_left: typing.Optional[int] = None,
face_padding_right: typing.Optional[int] = None,
sadtalker_settings: typing.Optional[SadTalkerSettings] = None,
- selected_model: typing.Optional[LipsyncModels] = None,
+ selected_model: typing.Optional[LipsyncTtsRequestSelectedModel] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
) -> LipsyncTtsPageOutput:
@@ -7342,7 +7370,7 @@ async def lipsync_tts(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- tts_provider : typing.Optional[TextToSpeechProviders]
+ tts_provider : typing.Optional[LipsyncTtsRequestTtsProvider]
uberduck_voice_name : typing.Optional[str]
@@ -7392,7 +7420,7 @@ async def lipsync_tts(
sadtalker_settings : typing.Optional[SadTalkerSettings]
- selected_model : typing.Optional[LipsyncModels]
+ selected_model : typing.Optional[LipsyncTtsRequestSelectedModel]
settings : typing.Optional[RunSettings]
@@ -7516,7 +7544,7 @@ async def text_to_speech(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
- tts_provider: typing.Optional[TextToSpeechProviders] = OMIT,
+ tts_provider: typing.Optional[TextToSpeechPageRequestTtsProvider] = OMIT,
uberduck_voice_name: typing.Optional[str] = OMIT,
uberduck_speaking_rate: typing.Optional[float] = OMIT,
google_voice_name: typing.Optional[str] = OMIT,
@@ -7549,7 +7577,7 @@ async def text_to_speech(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- tts_provider : typing.Optional[TextToSpeechProviders]
+ tts_provider : typing.Optional[TextToSpeechPageRequestTtsProvider]
uberduck_voice_name : typing.Optional[str]
@@ -7699,10 +7727,10 @@ async def speech_recognition(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
- selected_model: typing.Optional[AsrModels] = None,
+ selected_model: typing.Optional[SpeechRecognitionRequestSelectedModel] = None,
language: typing.Optional[str] = None,
- translation_model: typing.Optional[TranslationModels] = None,
- output_format: typing.Optional[AsrOutputFormat] = None,
+ translation_model: typing.Optional[SpeechRecognitionRequestTranslationModel] = None,
+ output_format: typing.Optional[SpeechRecognitionRequestOutputFormat] = None,
google_translate_target: typing.Optional[str] = None,
translation_source: typing.Optional[str] = None,
translation_target: typing.Optional[str] = None,
@@ -7723,13 +7751,13 @@ async def speech_recognition(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- selected_model : typing.Optional[AsrModels]
+ selected_model : typing.Optional[SpeechRecognitionRequestSelectedModel]
language : typing.Optional[str]
- translation_model : typing.Optional[TranslationModels]
+ translation_model : typing.Optional[SpeechRecognitionRequestTranslationModel]
- output_format : typing.Optional[AsrOutputFormat]
+ output_format : typing.Optional[SpeechRecognitionRequestOutputFormat]
google_translate_target : typing.Optional[str]
use `translation_model` & `translation_target` instead.
@@ -7851,7 +7879,7 @@ async def text_to_music(
guidance_scale: typing.Optional[float] = OMIT,
seed: typing.Optional[int] = OMIT,
sd2upscaling: typing.Optional[bool] = OMIT,
- selected_models: typing.Optional[typing.Sequence[Text2AudioModels]] = OMIT,
+ selected_models: typing.Optional[typing.Sequence[typing.Literal["audio_ldm"]]] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> Text2AudioPageOutput:
@@ -7881,7 +7909,7 @@ async def text_to_music(
sd2upscaling : typing.Optional[bool]
- selected_models : typing.Optional[typing.Sequence[Text2AudioModels]]
+ selected_models : typing.Optional[typing.Sequence[typing.Literal["audio_ldm"]]]
settings : typing.Optional[RunSettings]
@@ -7986,7 +8014,7 @@ async def translate(
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
texts: typing.Optional[typing.List[str]] = None,
- selected_model: typing.Optional[TranslationModels] = None,
+ selected_model: typing.Optional[TranslateRequestSelectedModel] = None,
translation_source: typing.Optional[str] = None,
translation_target: typing.Optional[str] = None,
glossary_document: typing.Optional[core.File] = None,
@@ -8005,7 +8033,7 @@ async def translate(
texts : typing.Optional[typing.List[str]]
- selected_model : typing.Optional[TranslationModels]
+ selected_model : typing.Optional[TranslateRequestSelectedModel]
translation_source : typing.Optional[str]
@@ -8114,7 +8142,7 @@ async def remix_image(
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
text_prompt: typing.Optional[str] = None,
- selected_model: typing.Optional[ImageToImageModels] = None,
+ selected_model: typing.Optional[RemixImageRequestSelectedModel] = None,
selected_controlnet_model: typing.Optional[RemixImageRequestSelectedControlnetModel] = None,
negative_prompt: typing.Optional[str] = None,
num_outputs: typing.Optional[int] = None,
@@ -8144,7 +8172,7 @@ async def remix_image(
text_prompt : typing.Optional[str]
- selected_model : typing.Optional[ImageToImageModels]
+ selected_model : typing.Optional[RemixImageRequestSelectedModel]
selected_controlnet_model : typing.Optional[RemixImageRequestSelectedControlnetModel]
@@ -8286,8 +8314,8 @@ async def text_to_image(
guidance_scale: typing.Optional[float] = OMIT,
seed: typing.Optional[int] = OMIT,
sd2upscaling: typing.Optional[bool] = OMIT,
- selected_models: typing.Optional[typing.Sequence[TextToImageModels]] = OMIT,
- scheduler: typing.Optional[Schedulers] = OMIT,
+ selected_models: typing.Optional[typing.Sequence[CompareText2ImgPageRequestSelectedModelsItem]] = OMIT,
+ scheduler: typing.Optional[CompareText2ImgPageRequestScheduler] = OMIT,
edit_instruction: typing.Optional[str] = OMIT,
image_guidance_scale: typing.Optional[float] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
@@ -8325,9 +8353,9 @@ async def text_to_image(
sd2upscaling : typing.Optional[bool]
- selected_models : typing.Optional[typing.Sequence[TextToImageModels]]
+ selected_models : typing.Optional[typing.Sequence[CompareText2ImgPageRequestSelectedModelsItem]]
- scheduler : typing.Optional[Schedulers]
+ scheduler : typing.Optional[CompareText2ImgPageRequestScheduler]
edit_instruction : typing.Optional[str]
@@ -8447,7 +8475,7 @@ async def product_image(
obj_pos_x: typing.Optional[float] = None,
obj_pos_y: typing.Optional[float] = None,
mask_threshold: typing.Optional[float] = None,
- selected_model: typing.Optional[InpaintingModels] = None,
+ selected_model: typing.Optional[ProductImageRequestSelectedModel] = None,
negative_prompt: typing.Optional[str] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[int] = None,
@@ -8482,7 +8510,7 @@ async def product_image(
mask_threshold : typing.Optional[float]
- selected_model : typing.Optional[InpaintingModels]
+ selected_model : typing.Optional[ProductImageRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -8615,7 +8643,7 @@ async def portrait(
face_scale: typing.Optional[float] = None,
face_pos_x: typing.Optional[float] = None,
face_pos_y: typing.Optional[float] = None,
- selected_model: typing.Optional[InpaintingModels] = None,
+ selected_model: typing.Optional[PortraitRequestSelectedModel] = None,
negative_prompt: typing.Optional[str] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[int] = None,
@@ -8648,7 +8676,7 @@ async def portrait(
face_pos_y : typing.Optional[float]
- selected_model : typing.Optional[InpaintingModels]
+ selected_model : typing.Optional[PortraitRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -8781,7 +8809,7 @@ async def image_from_email(
face_scale: typing.Optional[float] = OMIT,
face_pos_x: typing.Optional[float] = OMIT,
face_pos_y: typing.Optional[float] = OMIT,
- selected_model: typing.Optional[InpaintingModels] = OMIT,
+ selected_model: typing.Optional[EmailFaceInpaintingPageRequestSelectedModel] = OMIT,
negative_prompt: typing.Optional[str] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[int] = OMIT,
@@ -8823,7 +8851,7 @@ async def image_from_email(
face_pos_y : typing.Optional[float]
- selected_model : typing.Optional[InpaintingModels]
+ selected_model : typing.Optional[EmailFaceInpaintingPageRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -8976,9 +9004,9 @@ async def image_from_web_search(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
- selected_model: typing.Optional[ImageToImageModels] = OMIT,
+ selected_model: typing.Optional[GoogleImageGenPageRequestSelectedModel] = OMIT,
negative_prompt: typing.Optional[str] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[int] = OMIT,
@@ -9004,12 +9032,12 @@ async def image_from_web_search(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
- selected_model : typing.Optional[ImageToImageModels]
+ selected_model : typing.Optional[GoogleImageGenPageRequestSelectedModel]
negative_prompt : typing.Optional[str]
@@ -9135,7 +9163,7 @@ async def remove_background(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.List[RecipeFunction]] = None,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
- selected_model: typing.Optional[ImageSegmentationModels] = None,
+ selected_model: typing.Optional[RemoveBackgroundRequestSelectedModel] = None,
mask_threshold: typing.Optional[float] = None,
rect_persepective_transform: typing.Optional[bool] = None,
reflection_opacity: typing.Optional[float] = None,
@@ -9158,7 +9186,7 @@ async def remove_background(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- selected_model : typing.Optional[ImageSegmentationModels]
+ selected_model : typing.Optional[RemoveBackgroundRequestSelectedModel]
mask_threshold : typing.Optional[float]
@@ -9276,7 +9304,7 @@ async def upscale(
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = None,
input_image: typing.Optional[core.File] = None,
input_video: typing.Optional[core.File] = None,
- selected_models: typing.Optional[typing.List[UpscalerModels]] = None,
+ selected_models: typing.Optional[typing.List[UpscaleRequestSelectedModelsItem]] = None,
selected_bg_model: typing.Optional[typing.Literal["real_esrgan_x2"]] = None,
settings: typing.Optional[RunSettings] = None,
request_options: typing.Optional[RequestOptions] = None,
@@ -9300,7 +9328,7 @@ async def upscale(
input_video : typing.Optional[core.File]
See core.File for more documentation
- selected_models : typing.Optional[typing.List[UpscalerModels]]
+ selected_models : typing.Optional[typing.List[UpscaleRequestSelectedModelsItem]]
selected_bg_model : typing.Optional[typing.Literal["real_esrgan_x2"]]
@@ -9405,7 +9433,7 @@ async def embed(
example_id: typing.Optional[str] = None,
functions: typing.Optional[typing.Sequence[RecipeFunction]] = OMIT,
variables: typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]] = OMIT,
- selected_model: typing.Optional[EmbeddingModels] = OMIT,
+ selected_model: typing.Optional[EmbeddingsPageRequestSelectedModel] = OMIT,
settings: typing.Optional[RunSettings] = OMIT,
request_options: typing.Optional[RequestOptions] = None,
) -> EmbeddingsPageOutput:
@@ -9421,7 +9449,7 @@ async def embed(
variables : typing.Optional[typing.Dict[str, typing.Optional[typing.Any]]]
Variables to be used as Jinja prompt templates and in functions as arguments
- selected_model : typing.Optional[EmbeddingModels]
+ selected_model : typing.Optional[EmbeddingsPageRequestSelectedModel]
settings : typing.Optional[RunSettings]
@@ -9525,19 +9553,19 @@ async def seo_people_also_ask_doc(
max_context_words: typing.Optional[int] = OMIT,
scroll_jump: typing.Optional[int] = OMIT,
doc_extract_url: typing.Optional[str] = OMIT,
- embedding_model: typing.Optional[EmbeddingModels] = OMIT,
+ embedding_model: typing.Optional[RelatedQnADocPageRequestEmbeddingModel] = OMIT,
dense_weight: typing.Optional[float] = OMIT,
task_instructions: typing.Optional[str] = OMIT,
query_instructions: typing.Optional[str] = OMIT,
selected_model: typing.Optional[LargeLanguageModels] = OMIT,
- citation_style: typing.Optional[CitationStyles] = OMIT,
+ citation_style: typing.Optional[RelatedQnADocPageRequestCitationStyle] = OMIT,
avoid_repetition: typing.Optional[bool] = OMIT,
num_outputs: typing.Optional[int] = OMIT,
quality: typing.Optional[float] = OMIT,
max_tokens: typing.Optional[int] = OMIT,
sampling_temperature: typing.Optional[float] = OMIT,
- response_format_type: typing.Optional[ResponseFormatType] = OMIT,
- serp_search_location: typing.Optional[SerpSearchLocations] = OMIT,
+ response_format_type: typing.Optional[RelatedQnADocPageRequestResponseFormatType] = OMIT,
+ serp_search_location: typing.Optional[SerpSearchLocation] = OMIT,
scaleserp_locations: typing.Optional[typing.Sequence[str]] = OMIT,
serp_search_type: typing.Optional[SerpSearchType] = OMIT,
scaleserp_search_field: typing.Optional[str] = OMIT,
@@ -9568,7 +9596,7 @@ async def seo_people_also_ask_doc(
doc_extract_url : typing.Optional[str]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[RelatedQnADocPageRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -9582,7 +9610,7 @@ async def seo_people_also_ask_doc(
selected_model : typing.Optional[LargeLanguageModels]
- citation_style : typing.Optional[CitationStyles]
+ citation_style : typing.Optional[RelatedQnADocPageRequestCitationStyle]
avoid_repetition : typing.Optional[bool]
@@ -9594,9 +9622,9 @@ async def seo_people_also_ask_doc(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[RelatedQnADocPageRequestResponseFormatType]
- serp_search_location : typing.Optional[SerpSearchLocations]
+ serp_search_location : typing.Optional[SerpSearchLocation]
scaleserp_locations : typing.Optional[typing.Sequence[str]]
DEPRECATED: use `serp_search_location` instead
diff --git a/src/gooey/copilot/__init__.py b/src/gooey/copilot/__init__.py
index 0d11408..3234b31 100644
--- a/src/gooey/copilot/__init__.py
+++ b/src/gooey/copilot/__init__.py
@@ -1,19 +1,33 @@
# This file was auto-generated by Fern from our API Definition.
from .types import (
+ CopilotCompletionRequestAsrModel,
+ CopilotCompletionRequestCitationStyle,
+ CopilotCompletionRequestEmbeddingModel,
CopilotCompletionRequestFunctionsItem,
CopilotCompletionRequestFunctionsItemTrigger,
+ CopilotCompletionRequestLipsyncModel,
CopilotCompletionRequestOpenaiTtsModel,
CopilotCompletionRequestOpenaiVoiceName,
+ CopilotCompletionRequestResponseFormatType,
CopilotCompletionRequestSadtalkerSettings,
CopilotCompletionRequestSadtalkerSettingsPreprocess,
+ CopilotCompletionRequestTranslationModel,
+ CopilotCompletionRequestTtsProvider,
)
__all__ = [
+ "CopilotCompletionRequestAsrModel",
+ "CopilotCompletionRequestCitationStyle",
+ "CopilotCompletionRequestEmbeddingModel",
"CopilotCompletionRequestFunctionsItem",
"CopilotCompletionRequestFunctionsItemTrigger",
+ "CopilotCompletionRequestLipsyncModel",
"CopilotCompletionRequestOpenaiTtsModel",
"CopilotCompletionRequestOpenaiVoiceName",
+ "CopilotCompletionRequestResponseFormatType",
"CopilotCompletionRequestSadtalkerSettings",
"CopilotCompletionRequestSadtalkerSettingsPreprocess",
+ "CopilotCompletionRequestTranslationModel",
+ "CopilotCompletionRequestTtsProvider",
]
diff --git a/src/gooey/copilot/client.py b/src/gooey/copilot/client.py
index 247a892..9dcc465 100644
--- a/src/gooey/copilot/client.py
+++ b/src/gooey/copilot/client.py
@@ -6,14 +6,14 @@
from .. import core
from ..types.conversation_entry import ConversationEntry
from ..types.large_language_models import LargeLanguageModels
-from ..types.embedding_models import EmbeddingModels
-from ..types.citation_styles import CitationStyles
-from ..types.asr_models import AsrModels
-from ..types.translation_models import TranslationModels
-from ..types.lipsync_models import LipsyncModels
+from .types.copilot_completion_request_embedding_model import CopilotCompletionRequestEmbeddingModel
+from .types.copilot_completion_request_citation_style import CopilotCompletionRequestCitationStyle
+from .types.copilot_completion_request_asr_model import CopilotCompletionRequestAsrModel
+from .types.copilot_completion_request_translation_model import CopilotCompletionRequestTranslationModel
+from .types.copilot_completion_request_lipsync_model import CopilotCompletionRequestLipsyncModel
from ..types.llm_tools import LlmTools
-from ..types.response_format_type import ResponseFormatType
-from ..types.text_to_speech_providers import TextToSpeechProviders
+from .types.copilot_completion_request_response_format_type import CopilotCompletionRequestResponseFormatType
+from .types.copilot_completion_request_tts_provider import CopilotCompletionRequestTtsProvider
from .types.copilot_completion_request_openai_voice_name import CopilotCompletionRequestOpenaiVoiceName
from .types.copilot_completion_request_openai_tts_model import CopilotCompletionRequestOpenaiTtsModel
from .types.copilot_completion_request_sadtalker_settings import CopilotCompletionRequestSadtalkerSettings
@@ -60,25 +60,25 @@ def completion(
max_references: typing.Optional[int] = None,
max_context_words: typing.Optional[int] = None,
scroll_jump: typing.Optional[int] = None,
- embedding_model: typing.Optional[EmbeddingModels] = None,
+ embedding_model: typing.Optional[CopilotCompletionRequestEmbeddingModel] = None,
dense_weight: typing.Optional[float] = None,
- citation_style: typing.Optional[CitationStyles] = None,
+ citation_style: typing.Optional[CopilotCompletionRequestCitationStyle] = None,
use_url_shortener: typing.Optional[bool] = None,
- asr_model: typing.Optional[AsrModels] = None,
+ asr_model: typing.Optional[CopilotCompletionRequestAsrModel] = None,
asr_language: typing.Optional[str] = None,
- translation_model: typing.Optional[TranslationModels] = None,
+ translation_model: typing.Optional[CopilotCompletionRequestTranslationModel] = None,
user_language: typing.Optional[str] = None,
input_glossary_document: typing.Optional[core.File] = None,
output_glossary_document: typing.Optional[core.File] = None,
- lipsync_model: typing.Optional[LipsyncModels] = None,
+ lipsync_model: typing.Optional[CopilotCompletionRequestLipsyncModel] = None,
tools: typing.Optional[typing.List[LlmTools]] = None,
avoid_repetition: typing.Optional[bool] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[float] = None,
max_tokens: typing.Optional[int] = None,
sampling_temperature: typing.Optional[float] = None,
- response_format_type: typing.Optional[ResponseFormatType] = None,
- tts_provider: typing.Optional[TextToSpeechProviders] = None,
+ response_format_type: typing.Optional[CopilotCompletionRequestResponseFormatType] = None,
+ tts_provider: typing.Optional[CopilotCompletionRequestTtsProvider] = None,
uberduck_voice_name: typing.Optional[str] = None,
uberduck_speaking_rate: typing.Optional[float] = None,
google_voice_name: typing.Optional[str] = None,
@@ -152,7 +152,7 @@ def completion(
scroll_jump : typing.Optional[int]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[CopilotCompletionRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -160,17 +160,17 @@ def completion(
Generally speaking, dense embeddings excel at understanding the context of the query, whereas sparse vectors excel at keyword matches.
- citation_style : typing.Optional[CitationStyles]
+ citation_style : typing.Optional[CopilotCompletionRequestCitationStyle]
use_url_shortener : typing.Optional[bool]
- asr_model : typing.Optional[AsrModels]
+ asr_model : typing.Optional[CopilotCompletionRequestAsrModel]
Choose a model to transcribe incoming audio messages to text.
asr_language : typing.Optional[str]
Choose a language to transcribe incoming audio messages to text.
- translation_model : typing.Optional[TranslationModels]
+ translation_model : typing.Optional[CopilotCompletionRequestTranslationModel]
user_language : typing.Optional[str]
Choose a language to translate incoming text & audio messages to English and responses back to your selected language. Useful for low-resource languages.
@@ -181,7 +181,7 @@ def completion(
output_glossary_document : typing.Optional[core.File]
See core.File for more documentation
- lipsync_model : typing.Optional[LipsyncModels]
+ lipsync_model : typing.Optional[CopilotCompletionRequestLipsyncModel]
tools : typing.Optional[typing.List[LlmTools]]
Give your copilot superpowers by giving it access to tools. Powered by [Function calling](https://platform.openai.com/docs/guides/function-calling).
@@ -196,9 +196,9 @@ def completion(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[CopilotCompletionRequestResponseFormatType]
- tts_provider : typing.Optional[TextToSpeechProviders]
+ tts_provider : typing.Optional[CopilotCompletionRequestTtsProvider]
uberduck_voice_name : typing.Optional[str]
@@ -412,25 +412,25 @@ async def completion(
max_references: typing.Optional[int] = None,
max_context_words: typing.Optional[int] = None,
scroll_jump: typing.Optional[int] = None,
- embedding_model: typing.Optional[EmbeddingModels] = None,
+ embedding_model: typing.Optional[CopilotCompletionRequestEmbeddingModel] = None,
dense_weight: typing.Optional[float] = None,
- citation_style: typing.Optional[CitationStyles] = None,
+ citation_style: typing.Optional[CopilotCompletionRequestCitationStyle] = None,
use_url_shortener: typing.Optional[bool] = None,
- asr_model: typing.Optional[AsrModels] = None,
+ asr_model: typing.Optional[CopilotCompletionRequestAsrModel] = None,
asr_language: typing.Optional[str] = None,
- translation_model: typing.Optional[TranslationModels] = None,
+ translation_model: typing.Optional[CopilotCompletionRequestTranslationModel] = None,
user_language: typing.Optional[str] = None,
input_glossary_document: typing.Optional[core.File] = None,
output_glossary_document: typing.Optional[core.File] = None,
- lipsync_model: typing.Optional[LipsyncModels] = None,
+ lipsync_model: typing.Optional[CopilotCompletionRequestLipsyncModel] = None,
tools: typing.Optional[typing.List[LlmTools]] = None,
avoid_repetition: typing.Optional[bool] = None,
num_outputs: typing.Optional[int] = None,
quality: typing.Optional[float] = None,
max_tokens: typing.Optional[int] = None,
sampling_temperature: typing.Optional[float] = None,
- response_format_type: typing.Optional[ResponseFormatType] = None,
- tts_provider: typing.Optional[TextToSpeechProviders] = None,
+ response_format_type: typing.Optional[CopilotCompletionRequestResponseFormatType] = None,
+ tts_provider: typing.Optional[CopilotCompletionRequestTtsProvider] = None,
uberduck_voice_name: typing.Optional[str] = None,
uberduck_speaking_rate: typing.Optional[float] = None,
google_voice_name: typing.Optional[str] = None,
@@ -504,7 +504,7 @@ async def completion(
scroll_jump : typing.Optional[int]
- embedding_model : typing.Optional[EmbeddingModels]
+ embedding_model : typing.Optional[CopilotCompletionRequestEmbeddingModel]
dense_weight : typing.Optional[float]
@@ -512,17 +512,17 @@ async def completion(
Generally speaking, dense embeddings excel at understanding the context of the query, whereas sparse vectors excel at keyword matches.
- citation_style : typing.Optional[CitationStyles]
+ citation_style : typing.Optional[CopilotCompletionRequestCitationStyle]
use_url_shortener : typing.Optional[bool]
- asr_model : typing.Optional[AsrModels]
+ asr_model : typing.Optional[CopilotCompletionRequestAsrModel]
Choose a model to transcribe incoming audio messages to text.
asr_language : typing.Optional[str]
Choose a language to transcribe incoming audio messages to text.
- translation_model : typing.Optional[TranslationModels]
+ translation_model : typing.Optional[CopilotCompletionRequestTranslationModel]
user_language : typing.Optional[str]
Choose a language to translate incoming text & audio messages to English and responses back to your selected language. Useful for low-resource languages.
@@ -533,7 +533,7 @@ async def completion(
output_glossary_document : typing.Optional[core.File]
See core.File for more documentation
- lipsync_model : typing.Optional[LipsyncModels]
+ lipsync_model : typing.Optional[CopilotCompletionRequestLipsyncModel]
tools : typing.Optional[typing.List[LlmTools]]
Give your copilot superpowers by giving it access to tools. Powered by [Function calling](https://platform.openai.com/docs/guides/function-calling).
@@ -548,9 +548,9 @@ async def completion(
sampling_temperature : typing.Optional[float]
- response_format_type : typing.Optional[ResponseFormatType]
+ response_format_type : typing.Optional[CopilotCompletionRequestResponseFormatType]
- tts_provider : typing.Optional[TextToSpeechProviders]
+ tts_provider : typing.Optional[CopilotCompletionRequestTtsProvider]
uberduck_voice_name : typing.Optional[str]
diff --git a/src/gooey/copilot/types/__init__.py b/src/gooey/copilot/types/__init__.py
index 2094b54..1cdf619 100644
--- a/src/gooey/copilot/types/__init__.py
+++ b/src/gooey/copilot/types/__init__.py
@@ -1,19 +1,33 @@
# This file was auto-generated by Fern from our API Definition.
+from .copilot_completion_request_asr_model import CopilotCompletionRequestAsrModel
+from .copilot_completion_request_citation_style import CopilotCompletionRequestCitationStyle
+from .copilot_completion_request_embedding_model import CopilotCompletionRequestEmbeddingModel
from .copilot_completion_request_functions_item import CopilotCompletionRequestFunctionsItem
from .copilot_completion_request_functions_item_trigger import CopilotCompletionRequestFunctionsItemTrigger
+from .copilot_completion_request_lipsync_model import CopilotCompletionRequestLipsyncModel
from .copilot_completion_request_openai_tts_model import CopilotCompletionRequestOpenaiTtsModel
from .copilot_completion_request_openai_voice_name import CopilotCompletionRequestOpenaiVoiceName
+from .copilot_completion_request_response_format_type import CopilotCompletionRequestResponseFormatType
from .copilot_completion_request_sadtalker_settings import CopilotCompletionRequestSadtalkerSettings
from .copilot_completion_request_sadtalker_settings_preprocess import (
CopilotCompletionRequestSadtalkerSettingsPreprocess,
)
+from .copilot_completion_request_translation_model import CopilotCompletionRequestTranslationModel
+from .copilot_completion_request_tts_provider import CopilotCompletionRequestTtsProvider
__all__ = [
+ "CopilotCompletionRequestAsrModel",
+ "CopilotCompletionRequestCitationStyle",
+ "CopilotCompletionRequestEmbeddingModel",
"CopilotCompletionRequestFunctionsItem",
"CopilotCompletionRequestFunctionsItemTrigger",
+ "CopilotCompletionRequestLipsyncModel",
"CopilotCompletionRequestOpenaiTtsModel",
"CopilotCompletionRequestOpenaiVoiceName",
+ "CopilotCompletionRequestResponseFormatType",
"CopilotCompletionRequestSadtalkerSettings",
"CopilotCompletionRequestSadtalkerSettingsPreprocess",
+ "CopilotCompletionRequestTranslationModel",
+ "CopilotCompletionRequestTtsProvider",
]
diff --git a/src/gooey/copilot/types/copilot_completion_request_asr_model.py b/src/gooey/copilot/types/copilot_completion_request_asr_model.py
new file mode 100644
index 0000000..65ae0f5
--- /dev/null
+++ b/src/gooey/copilot/types/copilot_completion_request_asr_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CopilotCompletionRequestAsrModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/copilot/types/copilot_completion_request_citation_style.py b/src/gooey/copilot/types/copilot_completion_request_citation_style.py
new file mode 100644
index 0000000..1bb273a
--- /dev/null
+++ b/src/gooey/copilot/types/copilot_completion_request_citation_style.py
@@ -0,0 +1,25 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CopilotCompletionRequestCitationStyle = typing.Union[
+ typing.Literal[
+ "number",
+ "title",
+ "url",
+ "symbol",
+ "markdown",
+ "html",
+ "slack_mrkdwn",
+ "plaintext",
+ "number_markdown",
+ "number_html",
+ "number_slack_mrkdwn",
+ "number_plaintext",
+ "symbol_markdown",
+ "symbol_html",
+ "symbol_slack_mrkdwn",
+ "symbol_plaintext",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/copilot/types/copilot_completion_request_embedding_model.py b/src/gooey/copilot/types/copilot_completion_request_embedding_model.py
new file mode 100644
index 0000000..4655801
--- /dev/null
+++ b/src/gooey/copilot/types/copilot_completion_request_embedding_model.py
@@ -0,0 +1,18 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CopilotCompletionRequestEmbeddingModel = typing.Union[
+ typing.Literal[
+ "openai_3_large",
+ "openai_3_small",
+ "openai_ada_2",
+ "e5_large_v2",
+ "e5_base_v2",
+ "multilingual_e5_base",
+ "multilingual_e5_large",
+ "gte_large",
+ "gte_base",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/copilot/types/copilot_completion_request_lipsync_model.py b/src/gooey/copilot/types/copilot_completion_request_lipsync_model.py
new file mode 100644
index 0000000..865bc4b
--- /dev/null
+++ b/src/gooey/copilot/types/copilot_completion_request_lipsync_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CopilotCompletionRequestLipsyncModel = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/copilot/types/copilot_completion_request_response_format_type.py b/src/gooey/copilot/types/copilot_completion_request_response_format_type.py
new file mode 100644
index 0000000..3c9dbb0
--- /dev/null
+++ b/src/gooey/copilot/types/copilot_completion_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CopilotCompletionRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/copilot/types/copilot_completion_request_translation_model.py b/src/gooey/copilot/types/copilot_completion_request_translation_model.py
new file mode 100644
index 0000000..10b0b5a
--- /dev/null
+++ b/src/gooey/copilot/types/copilot_completion_request_translation_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CopilotCompletionRequestTranslationModel = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/copilot/types/copilot_completion_request_tts_provider.py b/src/gooey/copilot/types/copilot_completion_request_tts_provider.py
new file mode 100644
index 0000000..4dec4b0
--- /dev/null
+++ b/src/gooey/copilot/types/copilot_completion_request_tts_provider.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CopilotCompletionRequestTtsProvider = typing.Union[
+ typing.Literal["GOOGLE_TTS", "ELEVEN_LABS", "UBERDUCK", "BARK", "AZURE_TTS", "OPEN_AI"], typing.Any
+]
diff --git a/src/gooey/core/client_wrapper.py b/src/gooey/core/client_wrapper.py
index c630ac9..9c28e89 100644
--- a/src/gooey/core/client_wrapper.py
+++ b/src/gooey/core/client_wrapper.py
@@ -22,7 +22,7 @@ def get_headers(self) -> typing.Dict[str, str]:
headers: typing.Dict[str, str] = {
"X-Fern-Language": "Python",
"X-Fern-SDK-Name": "gooeyai",
- "X-Fern-SDK-Version": "0.0.1-beta25",
+ "X-Fern-SDK-Version": "0.0.1-beta26",
}
headers["Authorization"] = f"Bearer {self._get_api_key()}"
return headers
diff --git a/src/gooey/types/__init__.py b/src/gooey/types/__init__.py
index dfe0d3c..9087b38 100644
--- a/src/gooey/types/__init__.py
+++ b/src/gooey/types/__init__.py
@@ -4,21 +4,22 @@
from .agg_function_function import AggFunctionFunction
from .agg_function_result import AggFunctionResult
from .agg_function_result_function import AggFunctionResultFunction
-from .animation_models import AnimationModels
from .animation_prompt import AnimationPrompt
from .asr_chunk import AsrChunk
-from .asr_models import AsrModels
-from .asr_output_format import AsrOutputFormat
from .asr_output_json import AsrOutputJson
from .asr_page_output import AsrPageOutput
from .asr_page_output_output_text_item import AsrPageOutputOutputTextItem
from .asr_page_request import AsrPageRequest
+from .asr_page_request_output_format import AsrPageRequestOutputFormat
+from .asr_page_request_selected_model import AsrPageRequestSelectedModel
+from .asr_page_request_translation_model import AsrPageRequestTranslationModel
from .asr_page_status_response import AsrPageStatusResponse
from .async_api_response_model_v3 import AsyncApiResponseModelV3
from .balance_response import BalanceResponse
from .bot_broadcast_filters import BotBroadcastFilters
from .bot_broadcast_request_model import BotBroadcastRequestModel
from .bulk_eval_page_output import BulkEvalPageOutput
+from .bulk_eval_page_request_response_format_type import BulkEvalPageRequestResponseFormatType
from .bulk_eval_page_status_response import BulkEvalPageStatusResponse
from .bulk_runner_page_output import BulkRunnerPageOutput
from .bulk_runner_page_request import BulkRunnerPageRequest
@@ -31,18 +32,19 @@
from .chyron_plant_page_output import ChyronPlantPageOutput
from .chyron_plant_page_request import ChyronPlantPageRequest
from .chyron_plant_page_status_response import ChyronPlantPageStatusResponse
-from .citation_styles import CitationStyles
-from .combine_documents_chains import CombineDocumentsChains
from .compare_llm_page_output import CompareLlmPageOutput
+from .compare_llm_page_request_response_format_type import CompareLlmPageRequestResponseFormatType
from .compare_llm_page_status_response import CompareLlmPageStatusResponse
from .compare_text2img_page_output import CompareText2ImgPageOutput
+from .compare_text2img_page_request_scheduler import CompareText2ImgPageRequestScheduler
+from .compare_text2img_page_request_selected_models_item import CompareText2ImgPageRequestSelectedModelsItem
from .compare_text2img_page_status_response import CompareText2ImgPageStatusResponse
from .compare_upscaler_page_output import CompareUpscalerPageOutput
from .compare_upscaler_page_request import CompareUpscalerPageRequest
+from .compare_upscaler_page_request_selected_models_item import CompareUpscalerPageRequestSelectedModelsItem
from .compare_upscaler_page_status_response import CompareUpscalerPageStatusResponse
from .console_logs import ConsoleLogs
from .console_logs_level import ConsoleLogsLevel
-from .control_net_models import ControlNetModels
from .conversation_entry import ConversationEntry
from .conversation_entry_content import ConversationEntryContent
from .conversation_entry_content_item import (
@@ -53,28 +55,47 @@
from .conversation_entry_role import ConversationEntryRole
from .conversation_start import ConversationStart
from .create_stream_request import CreateStreamRequest
+from .create_stream_request_asr_model import CreateStreamRequestAsrModel
+from .create_stream_request_citation_style import CreateStreamRequestCitationStyle
+from .create_stream_request_embedding_model import CreateStreamRequestEmbeddingModel
+from .create_stream_request_lipsync_model import CreateStreamRequestLipsyncModel
from .create_stream_request_openai_tts_model import CreateStreamRequestOpenaiTtsModel
from .create_stream_request_openai_voice_name import CreateStreamRequestOpenaiVoiceName
+from .create_stream_request_response_format_type import CreateStreamRequestResponseFormatType
+from .create_stream_request_translation_model import CreateStreamRequestTranslationModel
+from .create_stream_request_tts_provider import CreateStreamRequestTtsProvider
from .create_stream_response import CreateStreamResponse
from .deforum_sd_page_output import DeforumSdPageOutput
+from .deforum_sd_page_request_selected_model import DeforumSdPageRequestSelectedModel
from .deforum_sd_page_status_response import DeforumSdPageStatusResponse
from .doc_extract_page_output import DocExtractPageOutput
from .doc_extract_page_request import DocExtractPageRequest
+from .doc_extract_page_request_response_format_type import DocExtractPageRequestResponseFormatType
+from .doc_extract_page_request_selected_asr_model import DocExtractPageRequestSelectedAsrModel
from .doc_extract_page_status_response import DocExtractPageStatusResponse
from .doc_search_page_output import DocSearchPageOutput
+from .doc_search_page_request_citation_style import DocSearchPageRequestCitationStyle
+from .doc_search_page_request_embedding_model import DocSearchPageRequestEmbeddingModel
from .doc_search_page_request_keyword_query import DocSearchPageRequestKeywordQuery
+from .doc_search_page_request_response_format_type import DocSearchPageRequestResponseFormatType
from .doc_search_page_status_response import DocSearchPageStatusResponse
from .doc_summary_page_output import DocSummaryPageOutput
from .doc_summary_page_request import DocSummaryPageRequest
+from .doc_summary_page_request_response_format_type import DocSummaryPageRequestResponseFormatType
+from .doc_summary_page_request_selected_asr_model import DocSummaryPageRequestSelectedAsrModel
from .doc_summary_page_status_response import DocSummaryPageStatusResponse
+from .doc_summary_request_response_format_type import DocSummaryRequestResponseFormatType
+from .doc_summary_request_selected_asr_model import DocSummaryRequestSelectedAsrModel
from .email_face_inpainting_page_output import EmailFaceInpaintingPageOutput
+from .email_face_inpainting_page_request_selected_model import EmailFaceInpaintingPageRequestSelectedModel
from .email_face_inpainting_page_status_response import EmailFaceInpaintingPageStatusResponse
-from .embedding_models import EmbeddingModels
from .embeddings_page_output import EmbeddingsPageOutput
+from .embeddings_page_request_selected_model import EmbeddingsPageRequestSelectedModel
from .embeddings_page_status_response import EmbeddingsPageStatusResponse
from .eval_prompt import EvalPrompt
from .face_inpainting_page_output import FaceInpaintingPageOutput
from .face_inpainting_page_request import FaceInpaintingPageRequest
+from .face_inpainting_page_request_selected_model import FaceInpaintingPageRequestSelectedModel
from .face_inpainting_page_status_response import FaceInpaintingPageStatusResponse
from .final_response import FinalResponse
from .functions_page_output import FunctionsPageOutput
@@ -82,60 +103,90 @@
from .generic_error_response import GenericErrorResponse
from .generic_error_response_detail import GenericErrorResponseDetail
from .google_gpt_page_output import GoogleGptPageOutput
+from .google_gpt_page_request_embedding_model import GoogleGptPageRequestEmbeddingModel
+from .google_gpt_page_request_response_format_type import GoogleGptPageRequestResponseFormatType
from .google_gpt_page_status_response import GoogleGptPageStatusResponse
from .google_image_gen_page_output import GoogleImageGenPageOutput
+from .google_image_gen_page_request_selected_model import GoogleImageGenPageRequestSelectedModel
from .google_image_gen_page_status_response import GoogleImageGenPageStatusResponse
from .http_validation_error import HttpValidationError
-from .image_segmentation_models import ImageSegmentationModels
from .image_segmentation_page_output import ImageSegmentationPageOutput
from .image_segmentation_page_request import ImageSegmentationPageRequest
+from .image_segmentation_page_request_selected_model import ImageSegmentationPageRequestSelectedModel
from .image_segmentation_page_status_response import ImageSegmentationPageStatusResponse
-from .image_to_image_models import ImageToImageModels
from .image_url import ImageUrl
from .image_url_detail import ImageUrlDetail
from .img2img_page_output import Img2ImgPageOutput
from .img2img_page_request import Img2ImgPageRequest
from .img2img_page_request_selected_controlnet_model import Img2ImgPageRequestSelectedControlnetModel
+from .img2img_page_request_selected_controlnet_model_item import Img2ImgPageRequestSelectedControlnetModelItem
+from .img2img_page_request_selected_model import Img2ImgPageRequestSelectedModel
from .img2img_page_status_response import Img2ImgPageStatusResponse
-from .inpainting_models import InpaintingModels
from .large_language_models import LargeLanguageModels
from .letter_writer_page_output import LetterWriterPageOutput
from .letter_writer_page_request import LetterWriterPageRequest
from .letter_writer_page_status_response import LetterWriterPageStatusResponse
-from .lipsync_models import LipsyncModels
from .lipsync_page_output import LipsyncPageOutput
from .lipsync_page_request import LipsyncPageRequest
+from .lipsync_page_request_selected_model import LipsyncPageRequestSelectedModel
from .lipsync_page_status_response import LipsyncPageStatusResponse
+from .lipsync_request_selected_model import LipsyncRequestSelectedModel
from .lipsync_tts_page_output import LipsyncTtsPageOutput
from .lipsync_tts_page_request import LipsyncTtsPageRequest
from .lipsync_tts_page_request_openai_tts_model import LipsyncTtsPageRequestOpenaiTtsModel
from .lipsync_tts_page_request_openai_voice_name import LipsyncTtsPageRequestOpenaiVoiceName
+from .lipsync_tts_page_request_selected_model import LipsyncTtsPageRequestSelectedModel
+from .lipsync_tts_page_request_tts_provider import LipsyncTtsPageRequestTtsProvider
from .lipsync_tts_page_status_response import LipsyncTtsPageStatusResponse
from .lipsync_tts_request_openai_tts_model import LipsyncTtsRequestOpenaiTtsModel
from .lipsync_tts_request_openai_voice_name import LipsyncTtsRequestOpenaiVoiceName
+from .lipsync_tts_request_selected_model import LipsyncTtsRequestSelectedModel
+from .lipsync_tts_request_tts_provider import LipsyncTtsRequestTtsProvider
from .llm_tools import LlmTools
from .message_part import MessagePart
from .object_inpainting_page_output import ObjectInpaintingPageOutput
from .object_inpainting_page_request import ObjectInpaintingPageRequest
+from .object_inpainting_page_request_selected_model import ObjectInpaintingPageRequestSelectedModel
from .object_inpainting_page_status_response import ObjectInpaintingPageStatusResponse
+from .portrait_request_selected_model import PortraitRequestSelectedModel
+from .product_image_request_selected_model import ProductImageRequestSelectedModel
from .prompt_tree_node import PromptTreeNode
from .prompt_tree_node_prompt import PromptTreeNodePrompt
from .qr_code_generator_page_output import QrCodeGeneratorPageOutput
from .qr_code_generator_page_request import QrCodeGeneratorPageRequest
+from .qr_code_generator_page_request_image_prompt_controlnet_models_item import (
+ QrCodeGeneratorPageRequestImagePromptControlnetModelsItem,
+)
+from .qr_code_generator_page_request_scheduler import QrCodeGeneratorPageRequestScheduler
+from .qr_code_generator_page_request_selected_controlnet_model_item import (
+ QrCodeGeneratorPageRequestSelectedControlnetModelItem,
+)
+from .qr_code_generator_page_request_selected_model import QrCodeGeneratorPageRequestSelectedModel
from .qr_code_generator_page_status_response import QrCodeGeneratorPageStatusResponse
+from .qr_code_request_image_prompt_controlnet_models_item import QrCodeRequestImagePromptControlnetModelsItem
+from .qr_code_request_scheduler import QrCodeRequestScheduler
+from .qr_code_request_selected_controlnet_model_item import QrCodeRequestSelectedControlnetModelItem
+from .qr_code_request_selected_model import QrCodeRequestSelectedModel
from .recipe_function import RecipeFunction
from .recipe_function_trigger import RecipeFunctionTrigger
from .recipe_run_state import RecipeRunState
from .related_doc_search_response import RelatedDocSearchResponse
from .related_google_gpt_response import RelatedGoogleGptResponse
from .related_qn_a_doc_page_output import RelatedQnADocPageOutput
+from .related_qn_a_doc_page_request_citation_style import RelatedQnADocPageRequestCitationStyle
+from .related_qn_a_doc_page_request_embedding_model import RelatedQnADocPageRequestEmbeddingModel
from .related_qn_a_doc_page_request_keyword_query import RelatedQnADocPageRequestKeywordQuery
+from .related_qn_a_doc_page_request_response_format_type import RelatedQnADocPageRequestResponseFormatType
from .related_qn_a_doc_page_status_response import RelatedQnADocPageStatusResponse
from .related_qn_a_page_output import RelatedQnAPageOutput
+from .related_qn_a_page_request_embedding_model import RelatedQnAPageRequestEmbeddingModel
+from .related_qn_a_page_request_response_format_type import RelatedQnAPageRequestResponseFormatType
from .related_qn_a_page_status_response import RelatedQnAPageStatusResponse
from .remix_image_request_selected_controlnet_model import RemixImageRequestSelectedControlnetModel
+from .remix_image_request_selected_controlnet_model_item import RemixImageRequestSelectedControlnetModelItem
+from .remix_image_request_selected_model import RemixImageRequestSelectedModel
+from .remove_background_request_selected_model import RemoveBackgroundRequestSelectedModel
from .reply_button import ReplyButton
-from .response_format_type import ResponseFormatType
from .response_model import ResponseModel
from .response_model_final_keyword_query import ResponseModelFinalKeywordQuery
from .response_model_final_prompt import ResponseModelFinalPrompt
@@ -144,32 +195,38 @@
from .run_start import RunStart
from .sad_talker_settings import SadTalkerSettings
from .sad_talker_settings_preprocess import SadTalkerSettingsPreprocess
-from .schedulers import Schedulers
from .search_reference import SearchReference
from .seo_summary_page_output import SeoSummaryPageOutput
+from .seo_summary_page_request_response_format_type import SeoSummaryPageRequestResponseFormatType
from .seo_summary_page_status_response import SeoSummaryPageStatusResponse
-from .serp_search_locations import SerpSearchLocations
+from .serp_search_location import SerpSearchLocation
from .serp_search_type import SerpSearchType
from .smart_gpt_page_output import SmartGptPageOutput
+from .smart_gpt_page_request_response_format_type import SmartGptPageRequestResponseFormatType
from .smart_gpt_page_status_response import SmartGptPageStatusResponse
from .social_lookup_email_page_output import SocialLookupEmailPageOutput
+from .social_lookup_email_page_request_response_format_type import SocialLookupEmailPageRequestResponseFormatType
from .social_lookup_email_page_status_response import SocialLookupEmailPageStatusResponse
+from .speech_recognition_request_output_format import SpeechRecognitionRequestOutputFormat
+from .speech_recognition_request_selected_model import SpeechRecognitionRequestSelectedModel
+from .speech_recognition_request_translation_model import SpeechRecognitionRequestTranslationModel
from .stream_error import StreamError
-from .text2audio_models import Text2AudioModels
+from .synthesize_data_request_response_format_type import SynthesizeDataRequestResponseFormatType
+from .synthesize_data_request_selected_asr_model import SynthesizeDataRequestSelectedAsrModel
from .text2audio_page_output import Text2AudioPageOutput
from .text2audio_page_status_response import Text2AudioPageStatusResponse
-from .text_to_image_models import TextToImageModels
from .text_to_speech_page_output import TextToSpeechPageOutput
from .text_to_speech_page_request_openai_tts_model import TextToSpeechPageRequestOpenaiTtsModel
from .text_to_speech_page_request_openai_voice_name import TextToSpeechPageRequestOpenaiVoiceName
+from .text_to_speech_page_request_tts_provider import TextToSpeechPageRequestTtsProvider
from .text_to_speech_page_status_response import TextToSpeechPageStatusResponse
-from .text_to_speech_providers import TextToSpeechProviders
from .training_data_model import TrainingDataModel
-from .translation_models import TranslationModels
+from .translate_request_selected_model import TranslateRequestSelectedModel
from .translation_page_output import TranslationPageOutput
from .translation_page_request import TranslationPageRequest
+from .translation_page_request_selected_model import TranslationPageRequestSelectedModel
from .translation_page_status_response import TranslationPageStatusResponse
-from .upscaler_models import UpscalerModels
+from .upscale_request_selected_models_item import UpscaleRequestSelectedModelsItem
from .validation_error import ValidationError
from .validation_error_loc_item import ValidationErrorLocItem
from .vcard import Vcard
@@ -177,12 +234,19 @@
from .video_bots_page_output_final_keyword_query import VideoBotsPageOutputFinalKeywordQuery
from .video_bots_page_output_final_prompt import VideoBotsPageOutputFinalPrompt
from .video_bots_page_request import VideoBotsPageRequest
+from .video_bots_page_request_asr_model import VideoBotsPageRequestAsrModel
+from .video_bots_page_request_citation_style import VideoBotsPageRequestCitationStyle
+from .video_bots_page_request_embedding_model import VideoBotsPageRequestEmbeddingModel
from .video_bots_page_request_functions_item import VideoBotsPageRequestFunctionsItem
from .video_bots_page_request_functions_item_trigger import VideoBotsPageRequestFunctionsItemTrigger
+from .video_bots_page_request_lipsync_model import VideoBotsPageRequestLipsyncModel
from .video_bots_page_request_openai_tts_model import VideoBotsPageRequestOpenaiTtsModel
from .video_bots_page_request_openai_voice_name import VideoBotsPageRequestOpenaiVoiceName
+from .video_bots_page_request_response_format_type import VideoBotsPageRequestResponseFormatType
from .video_bots_page_request_sadtalker_settings import VideoBotsPageRequestSadtalkerSettings
from .video_bots_page_request_sadtalker_settings_preprocess import VideoBotsPageRequestSadtalkerSettingsPreprocess
+from .video_bots_page_request_translation_model import VideoBotsPageRequestTranslationModel
+from .video_bots_page_request_tts_provider import VideoBotsPageRequestTtsProvider
from .video_bots_page_status_response import VideoBotsPageStatusResponse
__all__ = [
@@ -190,21 +254,22 @@
"AggFunctionFunction",
"AggFunctionResult",
"AggFunctionResultFunction",
- "AnimationModels",
"AnimationPrompt",
"AsrChunk",
- "AsrModels",
- "AsrOutputFormat",
"AsrOutputJson",
"AsrPageOutput",
"AsrPageOutputOutputTextItem",
"AsrPageRequest",
+ "AsrPageRequestOutputFormat",
+ "AsrPageRequestSelectedModel",
+ "AsrPageRequestTranslationModel",
"AsrPageStatusResponse",
"AsyncApiResponseModelV3",
"BalanceResponse",
"BotBroadcastFilters",
"BotBroadcastRequestModel",
"BulkEvalPageOutput",
+ "BulkEvalPageRequestResponseFormatType",
"BulkEvalPageStatusResponse",
"BulkRunnerPageOutput",
"BulkRunnerPageRequest",
@@ -217,18 +282,19 @@
"ChyronPlantPageOutput",
"ChyronPlantPageRequest",
"ChyronPlantPageStatusResponse",
- "CitationStyles",
- "CombineDocumentsChains",
"CompareLlmPageOutput",
+ "CompareLlmPageRequestResponseFormatType",
"CompareLlmPageStatusResponse",
"CompareText2ImgPageOutput",
+ "CompareText2ImgPageRequestScheduler",
+ "CompareText2ImgPageRequestSelectedModelsItem",
"CompareText2ImgPageStatusResponse",
"CompareUpscalerPageOutput",
"CompareUpscalerPageRequest",
+ "CompareUpscalerPageRequestSelectedModelsItem",
"CompareUpscalerPageStatusResponse",
"ConsoleLogs",
"ConsoleLogsLevel",
- "ControlNetModels",
"ConversationEntry",
"ConversationEntryContent",
"ConversationEntryContentItem",
@@ -237,28 +303,47 @@
"ConversationEntryRole",
"ConversationStart",
"CreateStreamRequest",
+ "CreateStreamRequestAsrModel",
+ "CreateStreamRequestCitationStyle",
+ "CreateStreamRequestEmbeddingModel",
+ "CreateStreamRequestLipsyncModel",
"CreateStreamRequestOpenaiTtsModel",
"CreateStreamRequestOpenaiVoiceName",
+ "CreateStreamRequestResponseFormatType",
+ "CreateStreamRequestTranslationModel",
+ "CreateStreamRequestTtsProvider",
"CreateStreamResponse",
"DeforumSdPageOutput",
+ "DeforumSdPageRequestSelectedModel",
"DeforumSdPageStatusResponse",
"DocExtractPageOutput",
"DocExtractPageRequest",
+ "DocExtractPageRequestResponseFormatType",
+ "DocExtractPageRequestSelectedAsrModel",
"DocExtractPageStatusResponse",
"DocSearchPageOutput",
+ "DocSearchPageRequestCitationStyle",
+ "DocSearchPageRequestEmbeddingModel",
"DocSearchPageRequestKeywordQuery",
+ "DocSearchPageRequestResponseFormatType",
"DocSearchPageStatusResponse",
"DocSummaryPageOutput",
"DocSummaryPageRequest",
+ "DocSummaryPageRequestResponseFormatType",
+ "DocSummaryPageRequestSelectedAsrModel",
"DocSummaryPageStatusResponse",
+ "DocSummaryRequestResponseFormatType",
+ "DocSummaryRequestSelectedAsrModel",
"EmailFaceInpaintingPageOutput",
+ "EmailFaceInpaintingPageRequestSelectedModel",
"EmailFaceInpaintingPageStatusResponse",
- "EmbeddingModels",
"EmbeddingsPageOutput",
+ "EmbeddingsPageRequestSelectedModel",
"EmbeddingsPageStatusResponse",
"EvalPrompt",
"FaceInpaintingPageOutput",
"FaceInpaintingPageRequest",
+ "FaceInpaintingPageRequestSelectedModel",
"FaceInpaintingPageStatusResponse",
"FinalResponse",
"FunctionsPageOutput",
@@ -266,60 +351,86 @@
"GenericErrorResponse",
"GenericErrorResponseDetail",
"GoogleGptPageOutput",
+ "GoogleGptPageRequestEmbeddingModel",
+ "GoogleGptPageRequestResponseFormatType",
"GoogleGptPageStatusResponse",
"GoogleImageGenPageOutput",
+ "GoogleImageGenPageRequestSelectedModel",
"GoogleImageGenPageStatusResponse",
"HttpValidationError",
- "ImageSegmentationModels",
"ImageSegmentationPageOutput",
"ImageSegmentationPageRequest",
+ "ImageSegmentationPageRequestSelectedModel",
"ImageSegmentationPageStatusResponse",
- "ImageToImageModels",
"ImageUrl",
"ImageUrlDetail",
"Img2ImgPageOutput",
"Img2ImgPageRequest",
"Img2ImgPageRequestSelectedControlnetModel",
+ "Img2ImgPageRequestSelectedControlnetModelItem",
+ "Img2ImgPageRequestSelectedModel",
"Img2ImgPageStatusResponse",
- "InpaintingModels",
"LargeLanguageModels",
"LetterWriterPageOutput",
"LetterWriterPageRequest",
"LetterWriterPageStatusResponse",
- "LipsyncModels",
"LipsyncPageOutput",
"LipsyncPageRequest",
+ "LipsyncPageRequestSelectedModel",
"LipsyncPageStatusResponse",
+ "LipsyncRequestSelectedModel",
"LipsyncTtsPageOutput",
"LipsyncTtsPageRequest",
"LipsyncTtsPageRequestOpenaiTtsModel",
"LipsyncTtsPageRequestOpenaiVoiceName",
+ "LipsyncTtsPageRequestSelectedModel",
+ "LipsyncTtsPageRequestTtsProvider",
"LipsyncTtsPageStatusResponse",
"LipsyncTtsRequestOpenaiTtsModel",
"LipsyncTtsRequestOpenaiVoiceName",
+ "LipsyncTtsRequestSelectedModel",
+ "LipsyncTtsRequestTtsProvider",
"LlmTools",
"MessagePart",
"ObjectInpaintingPageOutput",
"ObjectInpaintingPageRequest",
+ "ObjectInpaintingPageRequestSelectedModel",
"ObjectInpaintingPageStatusResponse",
+ "PortraitRequestSelectedModel",
+ "ProductImageRequestSelectedModel",
"PromptTreeNode",
"PromptTreeNodePrompt",
"QrCodeGeneratorPageOutput",
"QrCodeGeneratorPageRequest",
+ "QrCodeGeneratorPageRequestImagePromptControlnetModelsItem",
+ "QrCodeGeneratorPageRequestScheduler",
+ "QrCodeGeneratorPageRequestSelectedControlnetModelItem",
+ "QrCodeGeneratorPageRequestSelectedModel",
"QrCodeGeneratorPageStatusResponse",
+ "QrCodeRequestImagePromptControlnetModelsItem",
+ "QrCodeRequestScheduler",
+ "QrCodeRequestSelectedControlnetModelItem",
+ "QrCodeRequestSelectedModel",
"RecipeFunction",
"RecipeFunctionTrigger",
"RecipeRunState",
"RelatedDocSearchResponse",
"RelatedGoogleGptResponse",
"RelatedQnADocPageOutput",
+ "RelatedQnADocPageRequestCitationStyle",
+ "RelatedQnADocPageRequestEmbeddingModel",
"RelatedQnADocPageRequestKeywordQuery",
+ "RelatedQnADocPageRequestResponseFormatType",
"RelatedQnADocPageStatusResponse",
"RelatedQnAPageOutput",
+ "RelatedQnAPageRequestEmbeddingModel",
+ "RelatedQnAPageRequestResponseFormatType",
"RelatedQnAPageStatusResponse",
"RemixImageRequestSelectedControlnetModel",
+ "RemixImageRequestSelectedControlnetModelItem",
+ "RemixImageRequestSelectedModel",
+ "RemoveBackgroundRequestSelectedModel",
"ReplyButton",
- "ResponseFormatType",
"ResponseModel",
"ResponseModelFinalKeywordQuery",
"ResponseModelFinalPrompt",
@@ -328,32 +439,38 @@
"RunStart",
"SadTalkerSettings",
"SadTalkerSettingsPreprocess",
- "Schedulers",
"SearchReference",
"SeoSummaryPageOutput",
+ "SeoSummaryPageRequestResponseFormatType",
"SeoSummaryPageStatusResponse",
- "SerpSearchLocations",
+ "SerpSearchLocation",
"SerpSearchType",
"SmartGptPageOutput",
+ "SmartGptPageRequestResponseFormatType",
"SmartGptPageStatusResponse",
"SocialLookupEmailPageOutput",
+ "SocialLookupEmailPageRequestResponseFormatType",
"SocialLookupEmailPageStatusResponse",
+ "SpeechRecognitionRequestOutputFormat",
+ "SpeechRecognitionRequestSelectedModel",
+ "SpeechRecognitionRequestTranslationModel",
"StreamError",
- "Text2AudioModels",
+ "SynthesizeDataRequestResponseFormatType",
+ "SynthesizeDataRequestSelectedAsrModel",
"Text2AudioPageOutput",
"Text2AudioPageStatusResponse",
- "TextToImageModels",
"TextToSpeechPageOutput",
"TextToSpeechPageRequestOpenaiTtsModel",
"TextToSpeechPageRequestOpenaiVoiceName",
+ "TextToSpeechPageRequestTtsProvider",
"TextToSpeechPageStatusResponse",
- "TextToSpeechProviders",
"TrainingDataModel",
- "TranslationModels",
+ "TranslateRequestSelectedModel",
"TranslationPageOutput",
"TranslationPageRequest",
+ "TranslationPageRequestSelectedModel",
"TranslationPageStatusResponse",
- "UpscalerModels",
+ "UpscaleRequestSelectedModelsItem",
"ValidationError",
"ValidationErrorLocItem",
"Vcard",
@@ -361,11 +478,18 @@
"VideoBotsPageOutputFinalKeywordQuery",
"VideoBotsPageOutputFinalPrompt",
"VideoBotsPageRequest",
+ "VideoBotsPageRequestAsrModel",
+ "VideoBotsPageRequestCitationStyle",
+ "VideoBotsPageRequestEmbeddingModel",
"VideoBotsPageRequestFunctionsItem",
"VideoBotsPageRequestFunctionsItemTrigger",
+ "VideoBotsPageRequestLipsyncModel",
"VideoBotsPageRequestOpenaiTtsModel",
"VideoBotsPageRequestOpenaiVoiceName",
+ "VideoBotsPageRequestResponseFormatType",
"VideoBotsPageRequestSadtalkerSettings",
"VideoBotsPageRequestSadtalkerSettingsPreprocess",
+ "VideoBotsPageRequestTranslationModel",
+ "VideoBotsPageRequestTtsProvider",
"VideoBotsPageStatusResponse",
]
diff --git a/src/gooey/types/animation_models.py b/src/gooey/types/animation_models.py
deleted file mode 100644
index 8ad7a84..0000000
--- a/src/gooey/types/animation_models.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-AnimationModels = typing.Union[typing.Literal["protogen_2_2", "epicdream"], typing.Any]
diff --git a/src/gooey/types/asr_output_format.py b/src/gooey/types/asr_output_format.py
deleted file mode 100644
index b3b0e2d..0000000
--- a/src/gooey/types/asr_output_format.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-AsrOutputFormat = typing.Union[typing.Literal["text", "json", "srt", "vtt"], typing.Any]
diff --git a/src/gooey/types/asr_page_request.py b/src/gooey/types/asr_page_request.py
index 1210679..1d35181 100644
--- a/src/gooey/types/asr_page_request.py
+++ b/src/gooey/types/asr_page_request.py
@@ -4,9 +4,9 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .asr_models import AsrModels
-from .translation_models import TranslationModels
-from .asr_output_format import AsrOutputFormat
+from .asr_page_request_selected_model import AsrPageRequestSelectedModel
+from .asr_page_request_translation_model import AsrPageRequestTranslationModel
+from .asr_page_request_output_format import AsrPageRequestOutputFormat
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -19,10 +19,10 @@ class AsrPageRequest(UniversalBaseModel):
"""
documents: typing.List[str]
- selected_model: typing.Optional[AsrModels] = None
+ selected_model: typing.Optional[AsrPageRequestSelectedModel] = None
language: typing.Optional[str] = None
- translation_model: typing.Optional[TranslationModels] = None
- output_format: typing.Optional[AsrOutputFormat] = None
+ translation_model: typing.Optional[AsrPageRequestTranslationModel] = None
+ output_format: typing.Optional[AsrPageRequestOutputFormat] = None
google_translate_target: typing.Optional[str] = pydantic.Field(default=None)
"""
use `translation_model` & `translation_target` instead.
diff --git a/src/gooey/types/asr_page_request_output_format.py b/src/gooey/types/asr_page_request_output_format.py
new file mode 100644
index 0000000..101e681
--- /dev/null
+++ b/src/gooey/types/asr_page_request_output_format.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+AsrPageRequestOutputFormat = typing.Union[typing.Literal["text", "json", "srt", "vtt"], typing.Any]
diff --git a/src/gooey/types/asr_models.py b/src/gooey/types/asr_page_request_selected_model.py
similarity index 91%
rename from src/gooey/types/asr_models.py
rename to src/gooey/types/asr_page_request_selected_model.py
index 16c222a..4e80d3c 100644
--- a/src/gooey/types/asr_models.py
+++ b/src/gooey/types/asr_page_request_selected_model.py
@@ -2,7 +2,7 @@
import typing
-AsrModels = typing.Union[
+AsrPageRequestSelectedModel = typing.Union[
typing.Literal[
"whisper_large_v2",
"whisper_large_v3",
diff --git a/src/gooey/types/asr_page_request_translation_model.py b/src/gooey/types/asr_page_request_translation_model.py
new file mode 100644
index 0000000..d5dcef6
--- /dev/null
+++ b/src/gooey/types/asr_page_request_translation_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+AsrPageRequestTranslationModel = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/types/bulk_eval_page_request_response_format_type.py b/src/gooey/types/bulk_eval_page_request_response_format_type.py
new file mode 100644
index 0000000..f1c242f
--- /dev/null
+++ b/src/gooey/types/bulk_eval_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+BulkEvalPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/combine_documents_chains.py b/src/gooey/types/combine_documents_chains.py
deleted file mode 100644
index c457e00..0000000
--- a/src/gooey/types/combine_documents_chains.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-CombineDocumentsChains = typing.Literal["map_reduce"]
diff --git a/src/gooey/types/compare_llm_page_request_response_format_type.py b/src/gooey/types/compare_llm_page_request_response_format_type.py
new file mode 100644
index 0000000..a846068
--- /dev/null
+++ b/src/gooey/types/compare_llm_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CompareLlmPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/compare_text2img_page_request_scheduler.py b/src/gooey/types/compare_text2img_page_request_scheduler.py
new file mode 100644
index 0000000..29ce840
--- /dev/null
+++ b/src/gooey/types/compare_text2img_page_request_scheduler.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CompareText2ImgPageRequestScheduler = typing.Union[
+ typing.Literal[
+ "singlestep_dpm_solver",
+ "multistep_dpm_solver",
+ "dpm_sde",
+ "dpm_discrete",
+ "dpm_discrete_ancestral",
+ "unipc",
+ "lms_discrete",
+ "heun",
+ "euler",
+ "euler_ancestral",
+ "pndm",
+ "ddpm",
+ "ddim",
+ "deis",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/compare_text2img_page_request_selected_models_item.py b/src/gooey/types/compare_text2img_page_request_selected_models_item.py
new file mode 100644
index 0000000..4154491
--- /dev/null
+++ b/src/gooey/types/compare_text2img_page_request_selected_models_item.py
@@ -0,0 +1,22 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CompareText2ImgPageRequestSelectedModelsItem = typing.Union[
+ typing.Literal[
+ "dream_shaper",
+ "dreamlike_2",
+ "sd_2",
+ "sd_1_5",
+ "dall_e",
+ "dall_e_3",
+ "openjourney_2",
+ "openjourney",
+ "analog_diffusion",
+ "protogen_5_3",
+ "jack_qiao",
+ "rodent_diffusion_1_5",
+ "deepfloyd_if",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/compare_upscaler_page_request.py b/src/gooey/types/compare_upscaler_page_request.py
index 849d668..8cfb4e7 100644
--- a/src/gooey/types/compare_upscaler_page_request.py
+++ b/src/gooey/types/compare_upscaler_page_request.py
@@ -4,7 +4,7 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .upscaler_models import UpscalerModels
+from .compare_upscaler_page_request_selected_models_item import CompareUpscalerPageRequestSelectedModelsItem
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -23,7 +23,7 @@ class CompareUpscalerPageRequest(UniversalBaseModel):
The final upsampling scale of the image
"""
- selected_models: typing.Optional[typing.List[UpscalerModels]] = None
+ selected_models: typing.Optional[typing.List[CompareUpscalerPageRequestSelectedModelsItem]] = None
selected_bg_model: typing.Optional[typing.Literal["real_esrgan_x2"]] = None
settings: typing.Optional[RunSettings] = None
diff --git a/src/gooey/types/compare_upscaler_page_request_selected_models_item.py b/src/gooey/types/compare_upscaler_page_request_selected_models_item.py
new file mode 100644
index 0000000..eff4f6e
--- /dev/null
+++ b/src/gooey/types/compare_upscaler_page_request_selected_models_item.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CompareUpscalerPageRequestSelectedModelsItem = typing.Union[
+ typing.Literal["gfpgan_1_4", "real_esrgan_x2", "sd_x4", "real_esrgan", "gfpgan"], typing.Any
+]
diff --git a/src/gooey/types/create_stream_request.py b/src/gooey/types/create_stream_request.py
index bf4541f..2d4745b 100644
--- a/src/gooey/types/create_stream_request.py
+++ b/src/gooey/types/create_stream_request.py
@@ -7,14 +7,14 @@
from .recipe_function import RecipeFunction
from .conversation_entry import ConversationEntry
from .large_language_models import LargeLanguageModels
-from .embedding_models import EmbeddingModels
-from .citation_styles import CitationStyles
-from .asr_models import AsrModels
-from .translation_models import TranslationModels
-from .lipsync_models import LipsyncModels
+from .create_stream_request_embedding_model import CreateStreamRequestEmbeddingModel
+from .create_stream_request_citation_style import CreateStreamRequestCitationStyle
+from .create_stream_request_asr_model import CreateStreamRequestAsrModel
+from .create_stream_request_translation_model import CreateStreamRequestTranslationModel
+from .create_stream_request_lipsync_model import CreateStreamRequestLipsyncModel
from .llm_tools import LlmTools
-from .response_format_type import ResponseFormatType
-from .text_to_speech_providers import TextToSpeechProviders
+from .create_stream_request_response_format_type import CreateStreamRequestResponseFormatType
+from .create_stream_request_tts_provider import CreateStreamRequestTtsProvider
from .create_stream_request_openai_voice_name import CreateStreamRequestOpenaiVoiceName
from .create_stream_request_openai_tts_model import CreateStreamRequestOpenaiTtsModel
from .sad_talker_settings import SadTalkerSettings
@@ -85,16 +85,16 @@ class CreateStreamRequest(UniversalBaseModel):
max_references: typing.Optional[int] = None
max_context_words: typing.Optional[int] = None
scroll_jump: typing.Optional[int] = None
- embedding_model: typing.Optional[EmbeddingModels] = None
+ embedding_model: typing.Optional[CreateStreamRequestEmbeddingModel] = None
dense_weight: typing.Optional[float] = pydantic.Field(default=None)
"""
Weightage for dense vs sparse embeddings. `0` for sparse, `1` for dense, `0.5` for equal weight.
Generally speaking, dense embeddings excel at understanding the context of the query, whereas sparse vectors excel at keyword matches.
"""
- citation_style: typing.Optional[CitationStyles] = None
+ citation_style: typing.Optional[CreateStreamRequestCitationStyle] = None
use_url_shortener: typing.Optional[bool] = None
- asr_model: typing.Optional[AsrModels] = pydantic.Field(default=None)
+ asr_model: typing.Optional[CreateStreamRequestAsrModel] = pydantic.Field(default=None)
"""
Choose a model to transcribe incoming audio messages to text.
"""
@@ -104,7 +104,7 @@ class CreateStreamRequest(UniversalBaseModel):
Choose a language to transcribe incoming audio messages to text.
"""
- translation_model: typing.Optional[TranslationModels] = None
+ translation_model: typing.Optional[CreateStreamRequestTranslationModel] = None
user_language: typing.Optional[str] = pydantic.Field(default=None)
"""
Choose a language to translate incoming text & audio messages to English and responses back to your selected language. Useful for low-resource languages.
@@ -120,7 +120,7 @@ class CreateStreamRequest(UniversalBaseModel):
Translation Glossary for LLM Language (English) -> User Langauge
"""
- lipsync_model: typing.Optional[LipsyncModels] = None
+ lipsync_model: typing.Optional[CreateStreamRequestLipsyncModel] = None
tools: typing.Optional[typing.List[LlmTools]] = pydantic.Field(default=None)
"""
Give your copilot superpowers by giving it access to tools. Powered by [Function calling](https://platform.openai.com/docs/guides/function-calling).
@@ -131,8 +131,8 @@ class CreateStreamRequest(UniversalBaseModel):
quality: typing.Optional[float] = None
max_tokens: typing.Optional[int] = None
sampling_temperature: typing.Optional[float] = None
- response_format_type: typing.Optional[ResponseFormatType] = None
- tts_provider: typing.Optional[TextToSpeechProviders] = None
+ response_format_type: typing.Optional[CreateStreamRequestResponseFormatType] = None
+ tts_provider: typing.Optional[CreateStreamRequestTtsProvider] = None
uberduck_voice_name: typing.Optional[str] = None
uberduck_speaking_rate: typing.Optional[float] = None
google_voice_name: typing.Optional[str] = None
diff --git a/src/gooey/types/create_stream_request_asr_model.py b/src/gooey/types/create_stream_request_asr_model.py
new file mode 100644
index 0000000..af166fa
--- /dev/null
+++ b/src/gooey/types/create_stream_request_asr_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CreateStreamRequestAsrModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/citation_styles.py b/src/gooey/types/create_stream_request_citation_style.py
similarity index 90%
rename from src/gooey/types/citation_styles.py
rename to src/gooey/types/create_stream_request_citation_style.py
index 4d822c2..e57bab1 100644
--- a/src/gooey/types/citation_styles.py
+++ b/src/gooey/types/create_stream_request_citation_style.py
@@ -2,7 +2,7 @@
import typing
-CitationStyles = typing.Union[
+CreateStreamRequestCitationStyle = typing.Union[
typing.Literal[
"number",
"title",
diff --git a/src/gooey/types/embedding_models.py b/src/gooey/types/create_stream_request_embedding_model.py
similarity index 87%
rename from src/gooey/types/embedding_models.py
rename to src/gooey/types/create_stream_request_embedding_model.py
index 8007d2d..cef26bf 100644
--- a/src/gooey/types/embedding_models.py
+++ b/src/gooey/types/create_stream_request_embedding_model.py
@@ -2,7 +2,7 @@
import typing
-EmbeddingModels = typing.Union[
+CreateStreamRequestEmbeddingModel = typing.Union[
typing.Literal[
"openai_3_large",
"openai_3_small",
diff --git a/src/gooey/types/create_stream_request_lipsync_model.py b/src/gooey/types/create_stream_request_lipsync_model.py
new file mode 100644
index 0000000..c207d45
--- /dev/null
+++ b/src/gooey/types/create_stream_request_lipsync_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CreateStreamRequestLipsyncModel = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/types/create_stream_request_response_format_type.py b/src/gooey/types/create_stream_request_response_format_type.py
new file mode 100644
index 0000000..dc5024d
--- /dev/null
+++ b/src/gooey/types/create_stream_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CreateStreamRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/create_stream_request_translation_model.py b/src/gooey/types/create_stream_request_translation_model.py
new file mode 100644
index 0000000..3876937
--- /dev/null
+++ b/src/gooey/types/create_stream_request_translation_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CreateStreamRequestTranslationModel = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/types/create_stream_request_tts_provider.py b/src/gooey/types/create_stream_request_tts_provider.py
new file mode 100644
index 0000000..cad602d
--- /dev/null
+++ b/src/gooey/types/create_stream_request_tts_provider.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+CreateStreamRequestTtsProvider = typing.Union[
+ typing.Literal["GOOGLE_TTS", "ELEVEN_LABS", "UBERDUCK", "BARK", "AZURE_TTS", "OPEN_AI"], typing.Any
+]
diff --git a/src/gooey/types/deforum_sd_page_request_selected_model.py b/src/gooey/types/deforum_sd_page_request_selected_model.py
new file mode 100644
index 0000000..3af657a
--- /dev/null
+++ b/src/gooey/types/deforum_sd_page_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DeforumSdPageRequestSelectedModel = typing.Union[typing.Literal["protogen_2_2", "epicdream"], typing.Any]
diff --git a/src/gooey/types/doc_extract_page_request.py b/src/gooey/types/doc_extract_page_request.py
index 4beeb94..9690c6c 100644
--- a/src/gooey/types/doc_extract_page_request.py
+++ b/src/gooey/types/doc_extract_page_request.py
@@ -4,9 +4,9 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .asr_models import AsrModels
+from .doc_extract_page_request_selected_asr_model import DocExtractPageRequestSelectedAsrModel
from .large_language_models import LargeLanguageModels
-from .response_format_type import ResponseFormatType
+from .doc_extract_page_request_response_format_type import DocExtractPageRequestResponseFormatType
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -20,7 +20,7 @@ class DocExtractPageRequest(UniversalBaseModel):
documents: typing.List[str]
sheet_url: typing.Optional[str] = None
- selected_asr_model: typing.Optional[AsrModels] = None
+ selected_asr_model: typing.Optional[DocExtractPageRequestSelectedAsrModel] = None
google_translate_target: typing.Optional[str] = None
glossary_document: typing.Optional[str] = None
task_instructions: typing.Optional[str] = None
@@ -30,7 +30,7 @@ class DocExtractPageRequest(UniversalBaseModel):
quality: typing.Optional[float] = None
max_tokens: typing.Optional[int] = None
sampling_temperature: typing.Optional[float] = None
- response_format_type: typing.Optional[ResponseFormatType] = None
+ response_format_type: typing.Optional[DocExtractPageRequestResponseFormatType] = None
settings: typing.Optional[RunSettings] = None
if IS_PYDANTIC_V2:
diff --git a/src/gooey/types/doc_extract_page_request_response_format_type.py b/src/gooey/types/doc_extract_page_request_response_format_type.py
new file mode 100644
index 0000000..0ad7c14
--- /dev/null
+++ b/src/gooey/types/doc_extract_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocExtractPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/doc_extract_page_request_selected_asr_model.py b/src/gooey/types/doc_extract_page_request_selected_asr_model.py
new file mode 100644
index 0000000..a358400
--- /dev/null
+++ b/src/gooey/types/doc_extract_page_request_selected_asr_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocExtractPageRequestSelectedAsrModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/doc_search_page_request_citation_style.py b/src/gooey/types/doc_search_page_request_citation_style.py
new file mode 100644
index 0000000..b47b3be
--- /dev/null
+++ b/src/gooey/types/doc_search_page_request_citation_style.py
@@ -0,0 +1,25 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocSearchPageRequestCitationStyle = typing.Union[
+ typing.Literal[
+ "number",
+ "title",
+ "url",
+ "symbol",
+ "markdown",
+ "html",
+ "slack_mrkdwn",
+ "plaintext",
+ "number_markdown",
+ "number_html",
+ "number_slack_mrkdwn",
+ "number_plaintext",
+ "symbol_markdown",
+ "symbol_html",
+ "symbol_slack_mrkdwn",
+ "symbol_plaintext",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/doc_search_page_request_embedding_model.py b/src/gooey/types/doc_search_page_request_embedding_model.py
new file mode 100644
index 0000000..fb35612
--- /dev/null
+++ b/src/gooey/types/doc_search_page_request_embedding_model.py
@@ -0,0 +1,18 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocSearchPageRequestEmbeddingModel = typing.Union[
+ typing.Literal[
+ "openai_3_large",
+ "openai_3_small",
+ "openai_ada_2",
+ "e5_large_v2",
+ "e5_base_v2",
+ "multilingual_e5_base",
+ "multilingual_e5_large",
+ "gte_large",
+ "gte_base",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/doc_search_page_request_response_format_type.py b/src/gooey/types/doc_search_page_request_response_format_type.py
new file mode 100644
index 0000000..856b641
--- /dev/null
+++ b/src/gooey/types/doc_search_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocSearchPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/doc_summary_page_request.py b/src/gooey/types/doc_summary_page_request.py
index dadd11a..466ddc1 100644
--- a/src/gooey/types/doc_summary_page_request.py
+++ b/src/gooey/types/doc_summary_page_request.py
@@ -5,9 +5,8 @@
from .recipe_function import RecipeFunction
import pydantic
from .large_language_models import LargeLanguageModels
-from .combine_documents_chains import CombineDocumentsChains
-from .asr_models import AsrModels
-from .response_format_type import ResponseFormatType
+from .doc_summary_page_request_selected_asr_model import DocSummaryPageRequestSelectedAsrModel
+from .doc_summary_page_request_response_format_type import DocSummaryPageRequestResponseFormatType
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -23,15 +22,15 @@ class DocSummaryPageRequest(UniversalBaseModel):
task_instructions: typing.Optional[str] = None
merge_instructions: typing.Optional[str] = None
selected_model: typing.Optional[LargeLanguageModels] = None
- chain_type: typing.Optional[CombineDocumentsChains] = None
- selected_asr_model: typing.Optional[AsrModels] = None
+ chain_type: typing.Optional[typing.Literal["map_reduce"]] = None
+ selected_asr_model: typing.Optional[DocSummaryPageRequestSelectedAsrModel] = None
google_translate_target: typing.Optional[str] = None
avoid_repetition: typing.Optional[bool] = None
num_outputs: typing.Optional[int] = None
quality: typing.Optional[float] = None
max_tokens: typing.Optional[int] = None
sampling_temperature: typing.Optional[float] = None
- response_format_type: typing.Optional[ResponseFormatType] = None
+ response_format_type: typing.Optional[DocSummaryPageRequestResponseFormatType] = None
settings: typing.Optional[RunSettings] = None
if IS_PYDANTIC_V2:
diff --git a/src/gooey/types/doc_summary_page_request_response_format_type.py b/src/gooey/types/doc_summary_page_request_response_format_type.py
new file mode 100644
index 0000000..318ad7f
--- /dev/null
+++ b/src/gooey/types/doc_summary_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocSummaryPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/doc_summary_page_request_selected_asr_model.py b/src/gooey/types/doc_summary_page_request_selected_asr_model.py
new file mode 100644
index 0000000..c04cc7a
--- /dev/null
+++ b/src/gooey/types/doc_summary_page_request_selected_asr_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocSummaryPageRequestSelectedAsrModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/doc_summary_request_response_format_type.py b/src/gooey/types/doc_summary_request_response_format_type.py
new file mode 100644
index 0000000..8fabf9b
--- /dev/null
+++ b/src/gooey/types/doc_summary_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocSummaryRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/doc_summary_request_selected_asr_model.py b/src/gooey/types/doc_summary_request_selected_asr_model.py
new file mode 100644
index 0000000..8b8a338
--- /dev/null
+++ b/src/gooey/types/doc_summary_request_selected_asr_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+DocSummaryRequestSelectedAsrModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/email_face_inpainting_page_request_selected_model.py b/src/gooey/types/email_face_inpainting_page_request_selected_model.py
new file mode 100644
index 0000000..822b5a6
--- /dev/null
+++ b/src/gooey/types/email_face_inpainting_page_request_selected_model.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+EmailFaceInpaintingPageRequestSelectedModel = typing.Union[
+ typing.Literal["sd_2", "runway_ml", "dall_e", "jack_qiao"], typing.Any
+]
diff --git a/src/gooey/types/embeddings_page_request_selected_model.py b/src/gooey/types/embeddings_page_request_selected_model.py
new file mode 100644
index 0000000..a03ecc8
--- /dev/null
+++ b/src/gooey/types/embeddings_page_request_selected_model.py
@@ -0,0 +1,18 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+EmbeddingsPageRequestSelectedModel = typing.Union[
+ typing.Literal[
+ "openai_3_large",
+ "openai_3_small",
+ "openai_ada_2",
+ "e5_large_v2",
+ "e5_base_v2",
+ "multilingual_e5_base",
+ "multilingual_e5_large",
+ "gte_large",
+ "gte_base",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/face_inpainting_page_request.py b/src/gooey/types/face_inpainting_page_request.py
index 8e88a4f..a653205 100644
--- a/src/gooey/types/face_inpainting_page_request.py
+++ b/src/gooey/types/face_inpainting_page_request.py
@@ -4,7 +4,7 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .inpainting_models import InpaintingModels
+from .face_inpainting_page_request_selected_model import FaceInpaintingPageRequestSelectedModel
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -21,7 +21,7 @@ class FaceInpaintingPageRequest(UniversalBaseModel):
face_scale: typing.Optional[float] = None
face_pos_x: typing.Optional[float] = None
face_pos_y: typing.Optional[float] = None
- selected_model: typing.Optional[InpaintingModels] = None
+ selected_model: typing.Optional[FaceInpaintingPageRequestSelectedModel] = None
negative_prompt: typing.Optional[str] = None
num_outputs: typing.Optional[int] = None
quality: typing.Optional[int] = None
diff --git a/src/gooey/types/face_inpainting_page_request_selected_model.py b/src/gooey/types/face_inpainting_page_request_selected_model.py
new file mode 100644
index 0000000..9b8eab6
--- /dev/null
+++ b/src/gooey/types/face_inpainting_page_request_selected_model.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+FaceInpaintingPageRequestSelectedModel = typing.Union[
+ typing.Literal["sd_2", "runway_ml", "dall_e", "jack_qiao"], typing.Any
+]
diff --git a/src/gooey/types/google_gpt_page_request_embedding_model.py b/src/gooey/types/google_gpt_page_request_embedding_model.py
new file mode 100644
index 0000000..66f060f
--- /dev/null
+++ b/src/gooey/types/google_gpt_page_request_embedding_model.py
@@ -0,0 +1,18 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+GoogleGptPageRequestEmbeddingModel = typing.Union[
+ typing.Literal[
+ "openai_3_large",
+ "openai_3_small",
+ "openai_ada_2",
+ "e5_large_v2",
+ "e5_base_v2",
+ "multilingual_e5_base",
+ "multilingual_e5_large",
+ "gte_large",
+ "gte_base",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/google_gpt_page_request_response_format_type.py b/src/gooey/types/google_gpt_page_request_response_format_type.py
new file mode 100644
index 0000000..dd04dec
--- /dev/null
+++ b/src/gooey/types/google_gpt_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+GoogleGptPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/google_image_gen_page_request_selected_model.py b/src/gooey/types/google_image_gen_page_request_selected_model.py
new file mode 100644
index 0000000..c872962
--- /dev/null
+++ b/src/gooey/types/google_image_gen_page_request_selected_model.py
@@ -0,0 +1,21 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+GoogleImageGenPageRequestSelectedModel = typing.Union[
+ typing.Literal[
+ "dream_shaper",
+ "dreamlike_2",
+ "sd_2",
+ "sd_1_5",
+ "dall_e",
+ "instruct_pix2pix",
+ "openjourney_2",
+ "openjourney",
+ "analog_diffusion",
+ "protogen_5_3",
+ "jack_qiao",
+ "rodent_diffusion_1_5",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/image_segmentation_models.py b/src/gooey/types/image_segmentation_models.py
deleted file mode 100644
index aae4fee..0000000
--- a/src/gooey/types/image_segmentation_models.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-ImageSegmentationModels = typing.Union[typing.Literal["dis", "u2net"], typing.Any]
diff --git a/src/gooey/types/image_segmentation_page_request.py b/src/gooey/types/image_segmentation_page_request.py
index 3e1952c..a2ea60d 100644
--- a/src/gooey/types/image_segmentation_page_request.py
+++ b/src/gooey/types/image_segmentation_page_request.py
@@ -4,7 +4,7 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .image_segmentation_models import ImageSegmentationModels
+from .image_segmentation_page_request_selected_model import ImageSegmentationPageRequestSelectedModel
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -17,7 +17,7 @@ class ImageSegmentationPageRequest(UniversalBaseModel):
"""
input_image: str
- selected_model: typing.Optional[ImageSegmentationModels] = None
+ selected_model: typing.Optional[ImageSegmentationPageRequestSelectedModel] = None
mask_threshold: typing.Optional[float] = None
rect_persepective_transform: typing.Optional[bool] = None
reflection_opacity: typing.Optional[float] = None
diff --git a/src/gooey/types/image_segmentation_page_request_selected_model.py b/src/gooey/types/image_segmentation_page_request_selected_model.py
new file mode 100644
index 0000000..9b4b8d7
--- /dev/null
+++ b/src/gooey/types/image_segmentation_page_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+ImageSegmentationPageRequestSelectedModel = typing.Union[typing.Literal["dis", "u2net"], typing.Any]
diff --git a/src/gooey/types/img2img_page_request.py b/src/gooey/types/img2img_page_request.py
index 2b689aa..f3cfd2f 100644
--- a/src/gooey/types/img2img_page_request.py
+++ b/src/gooey/types/img2img_page_request.py
@@ -4,7 +4,7 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .image_to_image_models import ImageToImageModels
+from .img2img_page_request_selected_model import Img2ImgPageRequestSelectedModel
from .img2img_page_request_selected_controlnet_model import Img2ImgPageRequestSelectedControlnetModel
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -19,7 +19,7 @@ class Img2ImgPageRequest(UniversalBaseModel):
input_image: str
text_prompt: typing.Optional[str] = None
- selected_model: typing.Optional[ImageToImageModels] = None
+ selected_model: typing.Optional[Img2ImgPageRequestSelectedModel] = None
selected_controlnet_model: typing.Optional[Img2ImgPageRequestSelectedControlnetModel] = None
negative_prompt: typing.Optional[str] = None
num_outputs: typing.Optional[int] = None
diff --git a/src/gooey/types/img2img_page_request_selected_controlnet_model.py b/src/gooey/types/img2img_page_request_selected_controlnet_model.py
index 514d737..df9cb36 100644
--- a/src/gooey/types/img2img_page_request_selected_controlnet_model.py
+++ b/src/gooey/types/img2img_page_request_selected_controlnet_model.py
@@ -1,6 +1,19 @@
# This file was auto-generated by Fern from our API Definition.
import typing
-from .control_net_models import ControlNetModels
+from .img2img_page_request_selected_controlnet_model_item import Img2ImgPageRequestSelectedControlnetModelItem
-Img2ImgPageRequestSelectedControlnetModel = typing.Union[typing.List[ControlNetModels], ControlNetModels]
+Img2ImgPageRequestSelectedControlnetModel = typing.Union[
+ typing.List[Img2ImgPageRequestSelectedControlnetModelItem],
+ typing.Literal["sd_controlnet_canny"],
+ typing.Literal["sd_controlnet_depth"],
+ typing.Literal["sd_controlnet_hed"],
+ typing.Literal["sd_controlnet_mlsd"],
+ typing.Literal["sd_controlnet_normal"],
+ typing.Literal["sd_controlnet_openpose"],
+ typing.Literal["sd_controlnet_scribble"],
+ typing.Literal["sd_controlnet_seg"],
+ typing.Literal["sd_controlnet_tile"],
+ typing.Literal["sd_controlnet_brightness"],
+ typing.Literal["control_v1p_sd15_qrcode_monster_v2"],
+]
diff --git a/src/gooey/types/img2img_page_request_selected_controlnet_model_item.py b/src/gooey/types/img2img_page_request_selected_controlnet_model_item.py
new file mode 100644
index 0000000..1569cf5
--- /dev/null
+++ b/src/gooey/types/img2img_page_request_selected_controlnet_model_item.py
@@ -0,0 +1,20 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+Img2ImgPageRequestSelectedControlnetModelItem = typing.Union[
+ typing.Literal[
+ "sd_controlnet_canny",
+ "sd_controlnet_depth",
+ "sd_controlnet_hed",
+ "sd_controlnet_mlsd",
+ "sd_controlnet_normal",
+ "sd_controlnet_openpose",
+ "sd_controlnet_scribble",
+ "sd_controlnet_seg",
+ "sd_controlnet_tile",
+ "sd_controlnet_brightness",
+ "control_v1p_sd15_qrcode_monster_v2",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/img2img_page_request_selected_model.py b/src/gooey/types/img2img_page_request_selected_model.py
new file mode 100644
index 0000000..506c2b1
--- /dev/null
+++ b/src/gooey/types/img2img_page_request_selected_model.py
@@ -0,0 +1,21 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+Img2ImgPageRequestSelectedModel = typing.Union[
+ typing.Literal[
+ "dream_shaper",
+ "dreamlike_2",
+ "sd_2",
+ "sd_1_5",
+ "dall_e",
+ "instruct_pix2pix",
+ "openjourney_2",
+ "openjourney",
+ "analog_diffusion",
+ "protogen_5_3",
+ "jack_qiao",
+ "rodent_diffusion_1_5",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/inpainting_models.py b/src/gooey/types/inpainting_models.py
deleted file mode 100644
index f851858..0000000
--- a/src/gooey/types/inpainting_models.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-InpaintingModels = typing.Union[typing.Literal["sd_2", "runway_ml", "dall_e", "jack_qiao"], typing.Any]
diff --git a/src/gooey/types/lipsync_models.py b/src/gooey/types/lipsync_models.py
deleted file mode 100644
index 0ee41ee..0000000
--- a/src/gooey/types/lipsync_models.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-LipsyncModels = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/types/lipsync_page_request.py b/src/gooey/types/lipsync_page_request.py
index 5ea1d3f..2914a1e 100644
--- a/src/gooey/types/lipsync_page_request.py
+++ b/src/gooey/types/lipsync_page_request.py
@@ -5,7 +5,7 @@
from .recipe_function import RecipeFunction
import pydantic
from .sad_talker_settings import SadTalkerSettings
-from .lipsync_models import LipsyncModels
+from .lipsync_page_request_selected_model import LipsyncPageRequestSelectedModel
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -23,7 +23,7 @@ class LipsyncPageRequest(UniversalBaseModel):
face_padding_left: typing.Optional[int] = None
face_padding_right: typing.Optional[int] = None
sadtalker_settings: typing.Optional[SadTalkerSettings] = None
- selected_model: typing.Optional[LipsyncModels] = None
+ selected_model: typing.Optional[LipsyncPageRequestSelectedModel] = None
input_audio: typing.Optional[str] = None
settings: typing.Optional[RunSettings] = None
diff --git a/src/gooey/types/lipsync_page_request_selected_model.py b/src/gooey/types/lipsync_page_request_selected_model.py
new file mode 100644
index 0000000..da68ef8
--- /dev/null
+++ b/src/gooey/types/lipsync_page_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+LipsyncPageRequestSelectedModel = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/types/lipsync_request_selected_model.py b/src/gooey/types/lipsync_request_selected_model.py
new file mode 100644
index 0000000..c5614b4
--- /dev/null
+++ b/src/gooey/types/lipsync_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+LipsyncRequestSelectedModel = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/types/lipsync_tts_page_request.py b/src/gooey/types/lipsync_tts_page_request.py
index 9c90696..f4f5293 100644
--- a/src/gooey/types/lipsync_tts_page_request.py
+++ b/src/gooey/types/lipsync_tts_page_request.py
@@ -4,11 +4,11 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .text_to_speech_providers import TextToSpeechProviders
+from .lipsync_tts_page_request_tts_provider import LipsyncTtsPageRequestTtsProvider
from .lipsync_tts_page_request_openai_voice_name import LipsyncTtsPageRequestOpenaiVoiceName
from .lipsync_tts_page_request_openai_tts_model import LipsyncTtsPageRequestOpenaiTtsModel
from .sad_talker_settings import SadTalkerSettings
-from .lipsync_models import LipsyncModels
+from .lipsync_tts_page_request_selected_model import LipsyncTtsPageRequestSelectedModel
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -21,7 +21,7 @@ class LipsyncTtsPageRequest(UniversalBaseModel):
"""
text_prompt: str
- tts_provider: typing.Optional[TextToSpeechProviders] = None
+ tts_provider: typing.Optional[LipsyncTtsPageRequestTtsProvider] = None
uberduck_voice_name: typing.Optional[str] = None
uberduck_speaking_rate: typing.Optional[float] = None
google_voice_name: typing.Optional[str] = None
@@ -49,7 +49,7 @@ class LipsyncTtsPageRequest(UniversalBaseModel):
face_padding_left: typing.Optional[int] = None
face_padding_right: typing.Optional[int] = None
sadtalker_settings: typing.Optional[SadTalkerSettings] = None
- selected_model: typing.Optional[LipsyncModels] = None
+ selected_model: typing.Optional[LipsyncTtsPageRequestSelectedModel] = None
settings: typing.Optional[RunSettings] = None
if IS_PYDANTIC_V2:
diff --git a/src/gooey/types/lipsync_tts_page_request_selected_model.py b/src/gooey/types/lipsync_tts_page_request_selected_model.py
new file mode 100644
index 0000000..538058b
--- /dev/null
+++ b/src/gooey/types/lipsync_tts_page_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+LipsyncTtsPageRequestSelectedModel = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/types/lipsync_tts_page_request_tts_provider.py b/src/gooey/types/lipsync_tts_page_request_tts_provider.py
new file mode 100644
index 0000000..7e73fda
--- /dev/null
+++ b/src/gooey/types/lipsync_tts_page_request_tts_provider.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+LipsyncTtsPageRequestTtsProvider = typing.Union[
+ typing.Literal["GOOGLE_TTS", "ELEVEN_LABS", "UBERDUCK", "BARK", "AZURE_TTS", "OPEN_AI"], typing.Any
+]
diff --git a/src/gooey/types/lipsync_tts_request_selected_model.py b/src/gooey/types/lipsync_tts_request_selected_model.py
new file mode 100644
index 0000000..9ece5a9
--- /dev/null
+++ b/src/gooey/types/lipsync_tts_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+LipsyncTtsRequestSelectedModel = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/types/text_to_speech_providers.py b/src/gooey/types/lipsync_tts_request_tts_provider.py
similarity index 80%
rename from src/gooey/types/text_to_speech_providers.py
rename to src/gooey/types/lipsync_tts_request_tts_provider.py
index f86047f..1a23fe3 100644
--- a/src/gooey/types/text_to_speech_providers.py
+++ b/src/gooey/types/lipsync_tts_request_tts_provider.py
@@ -2,6 +2,6 @@
import typing
-TextToSpeechProviders = typing.Union[
+LipsyncTtsRequestTtsProvider = typing.Union[
typing.Literal["GOOGLE_TTS", "ELEVEN_LABS", "UBERDUCK", "BARK", "AZURE_TTS", "OPEN_AI"], typing.Any
]
diff --git a/src/gooey/types/object_inpainting_page_request.py b/src/gooey/types/object_inpainting_page_request.py
index 54d23b5..50b5b72 100644
--- a/src/gooey/types/object_inpainting_page_request.py
+++ b/src/gooey/types/object_inpainting_page_request.py
@@ -4,7 +4,7 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .inpainting_models import InpaintingModels
+from .object_inpainting_page_request_selected_model import ObjectInpaintingPageRequestSelectedModel
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -22,7 +22,7 @@ class ObjectInpaintingPageRequest(UniversalBaseModel):
obj_pos_x: typing.Optional[float] = None
obj_pos_y: typing.Optional[float] = None
mask_threshold: typing.Optional[float] = None
- selected_model: typing.Optional[InpaintingModels] = None
+ selected_model: typing.Optional[ObjectInpaintingPageRequestSelectedModel] = None
negative_prompt: typing.Optional[str] = None
num_outputs: typing.Optional[int] = None
quality: typing.Optional[int] = None
diff --git a/src/gooey/types/object_inpainting_page_request_selected_model.py b/src/gooey/types/object_inpainting_page_request_selected_model.py
new file mode 100644
index 0000000..92f1302
--- /dev/null
+++ b/src/gooey/types/object_inpainting_page_request_selected_model.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+ObjectInpaintingPageRequestSelectedModel = typing.Union[
+ typing.Literal["sd_2", "runway_ml", "dall_e", "jack_qiao"], typing.Any
+]
diff --git a/src/gooey/types/portrait_request_selected_model.py b/src/gooey/types/portrait_request_selected_model.py
new file mode 100644
index 0000000..6c4a5ce
--- /dev/null
+++ b/src/gooey/types/portrait_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+PortraitRequestSelectedModel = typing.Union[typing.Literal["sd_2", "runway_ml", "dall_e", "jack_qiao"], typing.Any]
diff --git a/src/gooey/types/product_image_request_selected_model.py b/src/gooey/types/product_image_request_selected_model.py
new file mode 100644
index 0000000..f1ce039
--- /dev/null
+++ b/src/gooey/types/product_image_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+ProductImageRequestSelectedModel = typing.Union[typing.Literal["sd_2", "runway_ml", "dall_e", "jack_qiao"], typing.Any]
diff --git a/src/gooey/types/qr_code_generator_page_request.py b/src/gooey/types/qr_code_generator_page_request.py
index 60831cb..68f3730 100644
--- a/src/gooey/types/qr_code_generator_page_request.py
+++ b/src/gooey/types/qr_code_generator_page_request.py
@@ -5,9 +5,14 @@
from .recipe_function import RecipeFunction
import pydantic
from .vcard import Vcard
-from .control_net_models import ControlNetModels
-from .text_to_image_models import TextToImageModels
-from .schedulers import Schedulers
+from .qr_code_generator_page_request_image_prompt_controlnet_models_item import (
+ QrCodeGeneratorPageRequestImagePromptControlnetModelsItem,
+)
+from .qr_code_generator_page_request_selected_model import QrCodeGeneratorPageRequestSelectedModel
+from .qr_code_generator_page_request_selected_controlnet_model_item import (
+ QrCodeGeneratorPageRequestSelectedControlnetModelItem,
+)
+from .qr_code_generator_page_request_scheduler import QrCodeGeneratorPageRequestScheduler
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -27,20 +32,24 @@ class QrCodeGeneratorPageRequest(UniversalBaseModel):
text_prompt: str
negative_prompt: typing.Optional[str] = None
image_prompt: typing.Optional[str] = None
- image_prompt_controlnet_models: typing.Optional[typing.List[ControlNetModels]] = None
+ image_prompt_controlnet_models: typing.Optional[
+ typing.List[QrCodeGeneratorPageRequestImagePromptControlnetModelsItem]
+ ] = None
image_prompt_strength: typing.Optional[float] = None
image_prompt_scale: typing.Optional[float] = None
image_prompt_pos_x: typing.Optional[float] = None
image_prompt_pos_y: typing.Optional[float] = None
- selected_model: typing.Optional[TextToImageModels] = None
- selected_controlnet_model: typing.Optional[typing.List[ControlNetModels]] = None
+ selected_model: typing.Optional[QrCodeGeneratorPageRequestSelectedModel] = None
+ selected_controlnet_model: typing.Optional[typing.List[QrCodeGeneratorPageRequestSelectedControlnetModelItem]] = (
+ None
+ )
output_width: typing.Optional[int] = None
output_height: typing.Optional[int] = None
guidance_scale: typing.Optional[float] = None
controlnet_conditioning_scale: typing.Optional[typing.List[float]] = None
num_outputs: typing.Optional[int] = None
quality: typing.Optional[int] = None
- scheduler: typing.Optional[Schedulers] = None
+ scheduler: typing.Optional[QrCodeGeneratorPageRequestScheduler] = None
seed: typing.Optional[int] = None
obj_scale: typing.Optional[float] = None
obj_pos_x: typing.Optional[float] = None
diff --git a/src/gooey/types/qr_code_generator_page_request_image_prompt_controlnet_models_item.py b/src/gooey/types/qr_code_generator_page_request_image_prompt_controlnet_models_item.py
new file mode 100644
index 0000000..508e7e9
--- /dev/null
+++ b/src/gooey/types/qr_code_generator_page_request_image_prompt_controlnet_models_item.py
@@ -0,0 +1,20 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+QrCodeGeneratorPageRequestImagePromptControlnetModelsItem = typing.Union[
+ typing.Literal[
+ "sd_controlnet_canny",
+ "sd_controlnet_depth",
+ "sd_controlnet_hed",
+ "sd_controlnet_mlsd",
+ "sd_controlnet_normal",
+ "sd_controlnet_openpose",
+ "sd_controlnet_scribble",
+ "sd_controlnet_seg",
+ "sd_controlnet_tile",
+ "sd_controlnet_brightness",
+ "control_v1p_sd15_qrcode_monster_v2",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/qr_code_generator_page_request_scheduler.py b/src/gooey/types/qr_code_generator_page_request_scheduler.py
new file mode 100644
index 0000000..e30308a
--- /dev/null
+++ b/src/gooey/types/qr_code_generator_page_request_scheduler.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+QrCodeGeneratorPageRequestScheduler = typing.Union[
+ typing.Literal[
+ "singlestep_dpm_solver",
+ "multistep_dpm_solver",
+ "dpm_sde",
+ "dpm_discrete",
+ "dpm_discrete_ancestral",
+ "unipc",
+ "lms_discrete",
+ "heun",
+ "euler",
+ "euler_ancestral",
+ "pndm",
+ "ddpm",
+ "ddim",
+ "deis",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/qr_code_generator_page_request_selected_controlnet_model_item.py b/src/gooey/types/qr_code_generator_page_request_selected_controlnet_model_item.py
new file mode 100644
index 0000000..c6f1967
--- /dev/null
+++ b/src/gooey/types/qr_code_generator_page_request_selected_controlnet_model_item.py
@@ -0,0 +1,20 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+QrCodeGeneratorPageRequestSelectedControlnetModelItem = typing.Union[
+ typing.Literal[
+ "sd_controlnet_canny",
+ "sd_controlnet_depth",
+ "sd_controlnet_hed",
+ "sd_controlnet_mlsd",
+ "sd_controlnet_normal",
+ "sd_controlnet_openpose",
+ "sd_controlnet_scribble",
+ "sd_controlnet_seg",
+ "sd_controlnet_tile",
+ "sd_controlnet_brightness",
+ "control_v1p_sd15_qrcode_monster_v2",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/qr_code_generator_page_request_selected_model.py b/src/gooey/types/qr_code_generator_page_request_selected_model.py
new file mode 100644
index 0000000..97282cb
--- /dev/null
+++ b/src/gooey/types/qr_code_generator_page_request_selected_model.py
@@ -0,0 +1,22 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+QrCodeGeneratorPageRequestSelectedModel = typing.Union[
+ typing.Literal[
+ "dream_shaper",
+ "dreamlike_2",
+ "sd_2",
+ "sd_1_5",
+ "dall_e",
+ "dall_e_3",
+ "openjourney_2",
+ "openjourney",
+ "analog_diffusion",
+ "protogen_5_3",
+ "jack_qiao",
+ "rodent_diffusion_1_5",
+ "deepfloyd_if",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/qr_code_request_image_prompt_controlnet_models_item.py b/src/gooey/types/qr_code_request_image_prompt_controlnet_models_item.py
new file mode 100644
index 0000000..3be2ab6
--- /dev/null
+++ b/src/gooey/types/qr_code_request_image_prompt_controlnet_models_item.py
@@ -0,0 +1,20 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+QrCodeRequestImagePromptControlnetModelsItem = typing.Union[
+ typing.Literal[
+ "sd_controlnet_canny",
+ "sd_controlnet_depth",
+ "sd_controlnet_hed",
+ "sd_controlnet_mlsd",
+ "sd_controlnet_normal",
+ "sd_controlnet_openpose",
+ "sd_controlnet_scribble",
+ "sd_controlnet_seg",
+ "sd_controlnet_tile",
+ "sd_controlnet_brightness",
+ "control_v1p_sd15_qrcode_monster_v2",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/schedulers.py b/src/gooey/types/qr_code_request_scheduler.py
similarity index 91%
rename from src/gooey/types/schedulers.py
rename to src/gooey/types/qr_code_request_scheduler.py
index d3b5398..890b204 100644
--- a/src/gooey/types/schedulers.py
+++ b/src/gooey/types/qr_code_request_scheduler.py
@@ -2,7 +2,7 @@
import typing
-Schedulers = typing.Union[
+QrCodeRequestScheduler = typing.Union[
typing.Literal[
"singlestep_dpm_solver",
"multistep_dpm_solver",
diff --git a/src/gooey/types/control_net_models.py b/src/gooey/types/qr_code_request_selected_controlnet_model_item.py
similarity index 89%
rename from src/gooey/types/control_net_models.py
rename to src/gooey/types/qr_code_request_selected_controlnet_model_item.py
index 5c5f68a..c5cdc8d 100644
--- a/src/gooey/types/control_net_models.py
+++ b/src/gooey/types/qr_code_request_selected_controlnet_model_item.py
@@ -2,7 +2,7 @@
import typing
-ControlNetModels = typing.Union[
+QrCodeRequestSelectedControlnetModelItem = typing.Union[
typing.Literal[
"sd_controlnet_canny",
"sd_controlnet_depth",
diff --git a/src/gooey/types/text_to_image_models.py b/src/gooey/types/qr_code_request_selected_model.py
similarity index 90%
rename from src/gooey/types/text_to_image_models.py
rename to src/gooey/types/qr_code_request_selected_model.py
index fd17514..7ea963c 100644
--- a/src/gooey/types/text_to_image_models.py
+++ b/src/gooey/types/qr_code_request_selected_model.py
@@ -2,7 +2,7 @@
import typing
-TextToImageModels = typing.Union[
+QrCodeRequestSelectedModel = typing.Union[
typing.Literal[
"dream_shaper",
"dreamlike_2",
diff --git a/src/gooey/types/related_qn_a_doc_page_request_citation_style.py b/src/gooey/types/related_qn_a_doc_page_request_citation_style.py
new file mode 100644
index 0000000..b98f002
--- /dev/null
+++ b/src/gooey/types/related_qn_a_doc_page_request_citation_style.py
@@ -0,0 +1,25 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+RelatedQnADocPageRequestCitationStyle = typing.Union[
+ typing.Literal[
+ "number",
+ "title",
+ "url",
+ "symbol",
+ "markdown",
+ "html",
+ "slack_mrkdwn",
+ "plaintext",
+ "number_markdown",
+ "number_html",
+ "number_slack_mrkdwn",
+ "number_plaintext",
+ "symbol_markdown",
+ "symbol_html",
+ "symbol_slack_mrkdwn",
+ "symbol_plaintext",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/related_qn_a_doc_page_request_embedding_model.py b/src/gooey/types/related_qn_a_doc_page_request_embedding_model.py
new file mode 100644
index 0000000..680bbb5
--- /dev/null
+++ b/src/gooey/types/related_qn_a_doc_page_request_embedding_model.py
@@ -0,0 +1,18 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+RelatedQnADocPageRequestEmbeddingModel = typing.Union[
+ typing.Literal[
+ "openai_3_large",
+ "openai_3_small",
+ "openai_ada_2",
+ "e5_large_v2",
+ "e5_base_v2",
+ "multilingual_e5_base",
+ "multilingual_e5_large",
+ "gte_large",
+ "gte_base",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/related_qn_a_doc_page_request_response_format_type.py b/src/gooey/types/related_qn_a_doc_page_request_response_format_type.py
new file mode 100644
index 0000000..c65a896
--- /dev/null
+++ b/src/gooey/types/related_qn_a_doc_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+RelatedQnADocPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/related_qn_a_page_request_embedding_model.py b/src/gooey/types/related_qn_a_page_request_embedding_model.py
new file mode 100644
index 0000000..a591920
--- /dev/null
+++ b/src/gooey/types/related_qn_a_page_request_embedding_model.py
@@ -0,0 +1,18 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+RelatedQnAPageRequestEmbeddingModel = typing.Union[
+ typing.Literal[
+ "openai_3_large",
+ "openai_3_small",
+ "openai_ada_2",
+ "e5_large_v2",
+ "e5_base_v2",
+ "multilingual_e5_base",
+ "multilingual_e5_large",
+ "gte_large",
+ "gte_base",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/related_qn_a_page_request_response_format_type.py b/src/gooey/types/related_qn_a_page_request_response_format_type.py
new file mode 100644
index 0000000..7bada87
--- /dev/null
+++ b/src/gooey/types/related_qn_a_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+RelatedQnAPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/remix_image_request_selected_controlnet_model.py b/src/gooey/types/remix_image_request_selected_controlnet_model.py
index 521d9c8..eea207f 100644
--- a/src/gooey/types/remix_image_request_selected_controlnet_model.py
+++ b/src/gooey/types/remix_image_request_selected_controlnet_model.py
@@ -1,6 +1,19 @@
# This file was auto-generated by Fern from our API Definition.
import typing
-from .control_net_models import ControlNetModels
+from .remix_image_request_selected_controlnet_model_item import RemixImageRequestSelectedControlnetModelItem
-RemixImageRequestSelectedControlnetModel = typing.Union[typing.List[ControlNetModels], ControlNetModels]
+RemixImageRequestSelectedControlnetModel = typing.Union[
+ typing.List[RemixImageRequestSelectedControlnetModelItem],
+ typing.Literal["sd_controlnet_canny"],
+ typing.Literal["sd_controlnet_depth"],
+ typing.Literal["sd_controlnet_hed"],
+ typing.Literal["sd_controlnet_mlsd"],
+ typing.Literal["sd_controlnet_normal"],
+ typing.Literal["sd_controlnet_openpose"],
+ typing.Literal["sd_controlnet_scribble"],
+ typing.Literal["sd_controlnet_seg"],
+ typing.Literal["sd_controlnet_tile"],
+ typing.Literal["sd_controlnet_brightness"],
+ typing.Literal["control_v1p_sd15_qrcode_monster_v2"],
+]
diff --git a/src/gooey/types/remix_image_request_selected_controlnet_model_item.py b/src/gooey/types/remix_image_request_selected_controlnet_model_item.py
new file mode 100644
index 0000000..b4f3ff0
--- /dev/null
+++ b/src/gooey/types/remix_image_request_selected_controlnet_model_item.py
@@ -0,0 +1,20 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+RemixImageRequestSelectedControlnetModelItem = typing.Union[
+ typing.Literal[
+ "sd_controlnet_canny",
+ "sd_controlnet_depth",
+ "sd_controlnet_hed",
+ "sd_controlnet_mlsd",
+ "sd_controlnet_normal",
+ "sd_controlnet_openpose",
+ "sd_controlnet_scribble",
+ "sd_controlnet_seg",
+ "sd_controlnet_tile",
+ "sd_controlnet_brightness",
+ "control_v1p_sd15_qrcode_monster_v2",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/image_to_image_models.py b/src/gooey/types/remix_image_request_selected_model.py
similarity index 89%
rename from src/gooey/types/image_to_image_models.py
rename to src/gooey/types/remix_image_request_selected_model.py
index 70c9201..245d6b0 100644
--- a/src/gooey/types/image_to_image_models.py
+++ b/src/gooey/types/remix_image_request_selected_model.py
@@ -2,7 +2,7 @@
import typing
-ImageToImageModels = typing.Union[
+RemixImageRequestSelectedModel = typing.Union[
typing.Literal[
"dream_shaper",
"dreamlike_2",
diff --git a/src/gooey/types/remove_background_request_selected_model.py b/src/gooey/types/remove_background_request_selected_model.py
new file mode 100644
index 0000000..c84f0e7
--- /dev/null
+++ b/src/gooey/types/remove_background_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+RemoveBackgroundRequestSelectedModel = typing.Union[typing.Literal["dis", "u2net"], typing.Any]
diff --git a/src/gooey/types/response_format_type.py b/src/gooey/types/response_format_type.py
deleted file mode 100644
index f8216e9..0000000
--- a/src/gooey/types/response_format_type.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-ResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/seo_summary_page_request_response_format_type.py b/src/gooey/types/seo_summary_page_request_response_format_type.py
new file mode 100644
index 0000000..26f948b
--- /dev/null
+++ b/src/gooey/types/seo_summary_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SeoSummaryPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/serp_search_locations.py b/src/gooey/types/serp_search_location.py
similarity index 98%
rename from src/gooey/types/serp_search_locations.py
rename to src/gooey/types/serp_search_location.py
index 2d5144d..9b64ad9 100644
--- a/src/gooey/types/serp_search_locations.py
+++ b/src/gooey/types/serp_search_location.py
@@ -2,7 +2,7 @@
import typing
-SerpSearchLocations = typing.Union[
+SerpSearchLocation = typing.Union[
typing.Literal[
"af",
"al",
diff --git a/src/gooey/types/smart_gpt_page_request_response_format_type.py b/src/gooey/types/smart_gpt_page_request_response_format_type.py
new file mode 100644
index 0000000..1eaf901
--- /dev/null
+++ b/src/gooey/types/smart_gpt_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SmartGptPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/social_lookup_email_page_request_response_format_type.py b/src/gooey/types/social_lookup_email_page_request_response_format_type.py
new file mode 100644
index 0000000..46c50db
--- /dev/null
+++ b/src/gooey/types/social_lookup_email_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SocialLookupEmailPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/speech_recognition_request_output_format.py b/src/gooey/types/speech_recognition_request_output_format.py
new file mode 100644
index 0000000..4d2cf2b
--- /dev/null
+++ b/src/gooey/types/speech_recognition_request_output_format.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SpeechRecognitionRequestOutputFormat = typing.Union[typing.Literal["text", "json", "srt", "vtt"], typing.Any]
diff --git a/src/gooey/types/speech_recognition_request_selected_model.py b/src/gooey/types/speech_recognition_request_selected_model.py
new file mode 100644
index 0000000..9d2d28f
--- /dev/null
+++ b/src/gooey/types/speech_recognition_request_selected_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SpeechRecognitionRequestSelectedModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/speech_recognition_request_translation_model.py b/src/gooey/types/speech_recognition_request_translation_model.py
new file mode 100644
index 0000000..886ab92
--- /dev/null
+++ b/src/gooey/types/speech_recognition_request_translation_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SpeechRecognitionRequestTranslationModel = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/types/synthesize_data_request_response_format_type.py b/src/gooey/types/synthesize_data_request_response_format_type.py
new file mode 100644
index 0000000..3ab37a9
--- /dev/null
+++ b/src/gooey/types/synthesize_data_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SynthesizeDataRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/synthesize_data_request_selected_asr_model.py b/src/gooey/types/synthesize_data_request_selected_asr_model.py
new file mode 100644
index 0000000..6c1bc21
--- /dev/null
+++ b/src/gooey/types/synthesize_data_request_selected_asr_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+SynthesizeDataRequestSelectedAsrModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/text2audio_models.py b/src/gooey/types/text2audio_models.py
deleted file mode 100644
index b3eb9eb..0000000
--- a/src/gooey/types/text2audio_models.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-Text2AudioModels = typing.Literal["audio_ldm"]
diff --git a/src/gooey/types/text_to_speech_page_request_tts_provider.py b/src/gooey/types/text_to_speech_page_request_tts_provider.py
new file mode 100644
index 0000000..a6b8938
--- /dev/null
+++ b/src/gooey/types/text_to_speech_page_request_tts_provider.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+TextToSpeechPageRequestTtsProvider = typing.Union[
+ typing.Literal["GOOGLE_TTS", "ELEVEN_LABS", "UBERDUCK", "BARK", "AZURE_TTS", "OPEN_AI"], typing.Any
+]
diff --git a/src/gooey/types/translate_request_selected_model.py b/src/gooey/types/translate_request_selected_model.py
new file mode 100644
index 0000000..b774b56
--- /dev/null
+++ b/src/gooey/types/translate_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+TranslateRequestSelectedModel = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/types/translation_models.py b/src/gooey/types/translation_models.py
deleted file mode 100644
index 136ecb8..0000000
--- a/src/gooey/types/translation_models.py
+++ /dev/null
@@ -1,5 +0,0 @@
-# This file was auto-generated by Fern from our API Definition.
-
-import typing
-
-TranslationModels = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/types/translation_page_request.py b/src/gooey/types/translation_page_request.py
index 6845f7f..9c033a6 100644
--- a/src/gooey/types/translation_page_request.py
+++ b/src/gooey/types/translation_page_request.py
@@ -4,7 +4,7 @@
import typing
from .recipe_function import RecipeFunction
import pydantic
-from .translation_models import TranslationModels
+from .translation_page_request_selected_model import TranslationPageRequestSelectedModel
from .run_settings import RunSettings
from ..core.pydantic_utilities import IS_PYDANTIC_V2
@@ -17,7 +17,7 @@ class TranslationPageRequest(UniversalBaseModel):
"""
texts: typing.Optional[typing.List[str]] = None
- selected_model: typing.Optional[TranslationModels] = None
+ selected_model: typing.Optional[TranslationPageRequestSelectedModel] = None
translation_source: typing.Optional[str] = None
translation_target: typing.Optional[str] = None
glossary_document: typing.Optional[str] = None
diff --git a/src/gooey/types/translation_page_request_selected_model.py b/src/gooey/types/translation_page_request_selected_model.py
new file mode 100644
index 0000000..62ae9ab
--- /dev/null
+++ b/src/gooey/types/translation_page_request_selected_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+TranslationPageRequestSelectedModel = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/types/upscaler_models.py b/src/gooey/types/upscale_request_selected_models_item.py
similarity index 78%
rename from src/gooey/types/upscaler_models.py
rename to src/gooey/types/upscale_request_selected_models_item.py
index 314c03a..1a8362e 100644
--- a/src/gooey/types/upscaler_models.py
+++ b/src/gooey/types/upscale_request_selected_models_item.py
@@ -2,6 +2,6 @@
import typing
-UpscalerModels = typing.Union[
+UpscaleRequestSelectedModelsItem = typing.Union[
typing.Literal["gfpgan_1_4", "real_esrgan_x2", "sd_x4", "real_esrgan", "gfpgan"], typing.Any
]
diff --git a/src/gooey/types/video_bots_page_request.py b/src/gooey/types/video_bots_page_request.py
index d1f4e31..6fb8b5e 100644
--- a/src/gooey/types/video_bots_page_request.py
+++ b/src/gooey/types/video_bots_page_request.py
@@ -6,14 +6,14 @@
import pydantic
from .conversation_entry import ConversationEntry
from .large_language_models import LargeLanguageModels
-from .embedding_models import EmbeddingModels
-from .citation_styles import CitationStyles
-from .asr_models import AsrModels
-from .translation_models import TranslationModels
-from .lipsync_models import LipsyncModels
+from .video_bots_page_request_embedding_model import VideoBotsPageRequestEmbeddingModel
+from .video_bots_page_request_citation_style import VideoBotsPageRequestCitationStyle
+from .video_bots_page_request_asr_model import VideoBotsPageRequestAsrModel
+from .video_bots_page_request_translation_model import VideoBotsPageRequestTranslationModel
+from .video_bots_page_request_lipsync_model import VideoBotsPageRequestLipsyncModel
from .llm_tools import LlmTools
-from .response_format_type import ResponseFormatType
-from .text_to_speech_providers import TextToSpeechProviders
+from .video_bots_page_request_response_format_type import VideoBotsPageRequestResponseFormatType
+from .video_bots_page_request_tts_provider import VideoBotsPageRequestTtsProvider
from .video_bots_page_request_openai_voice_name import VideoBotsPageRequestOpenaiVoiceName
from .video_bots_page_request_openai_tts_model import VideoBotsPageRequestOpenaiTtsModel
from .video_bots_page_request_sadtalker_settings import VideoBotsPageRequestSadtalkerSettings
@@ -52,16 +52,16 @@ class VideoBotsPageRequest(UniversalBaseModel):
max_references: typing.Optional[int] = None
max_context_words: typing.Optional[int] = None
scroll_jump: typing.Optional[int] = None
- embedding_model: typing.Optional[EmbeddingModels] = None
+ embedding_model: typing.Optional[VideoBotsPageRequestEmbeddingModel] = None
dense_weight: typing.Optional[float] = pydantic.Field(default=None)
"""
Weightage for dense vs sparse embeddings. `0` for sparse, `1` for dense, `0.5` for equal weight.
Generally speaking, dense embeddings excel at understanding the context of the query, whereas sparse vectors excel at keyword matches.
"""
- citation_style: typing.Optional[CitationStyles] = None
+ citation_style: typing.Optional[VideoBotsPageRequestCitationStyle] = None
use_url_shortener: typing.Optional[bool] = None
- asr_model: typing.Optional[AsrModels] = pydantic.Field(default=None)
+ asr_model: typing.Optional[VideoBotsPageRequestAsrModel] = pydantic.Field(default=None)
"""
Choose a model to transcribe incoming audio messages to text.
"""
@@ -71,7 +71,7 @@ class VideoBotsPageRequest(UniversalBaseModel):
Choose a language to transcribe incoming audio messages to text.
"""
- translation_model: typing.Optional[TranslationModels] = None
+ translation_model: typing.Optional[VideoBotsPageRequestTranslationModel] = None
user_language: typing.Optional[str] = pydantic.Field(default=None)
"""
Choose a language to translate incoming text & audio messages to English and responses back to your selected language. Useful for low-resource languages.
@@ -79,7 +79,7 @@ class VideoBotsPageRequest(UniversalBaseModel):
input_glossary_document: typing.Optional[str] = None
output_glossary_document: typing.Optional[str] = None
- lipsync_model: typing.Optional[LipsyncModels] = None
+ lipsync_model: typing.Optional[VideoBotsPageRequestLipsyncModel] = None
tools: typing.Optional[typing.List[LlmTools]] = pydantic.Field(default=None)
"""
Give your copilot superpowers by giving it access to tools. Powered by [Function calling](https://platform.openai.com/docs/guides/function-calling).
@@ -90,8 +90,8 @@ class VideoBotsPageRequest(UniversalBaseModel):
quality: typing.Optional[float] = None
max_tokens: typing.Optional[int] = None
sampling_temperature: typing.Optional[float] = None
- response_format_type: typing.Optional[ResponseFormatType] = None
- tts_provider: typing.Optional[TextToSpeechProviders] = None
+ response_format_type: typing.Optional[VideoBotsPageRequestResponseFormatType] = None
+ tts_provider: typing.Optional[VideoBotsPageRequestTtsProvider] = None
uberduck_voice_name: typing.Optional[str] = None
uberduck_speaking_rate: typing.Optional[float] = None
google_voice_name: typing.Optional[str] = None
diff --git a/src/gooey/types/video_bots_page_request_asr_model.py b/src/gooey/types/video_bots_page_request_asr_model.py
new file mode 100644
index 0000000..7db13bc
--- /dev/null
+++ b/src/gooey/types/video_bots_page_request_asr_model.py
@@ -0,0 +1,23 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+VideoBotsPageRequestAsrModel = typing.Union[
+ typing.Literal[
+ "whisper_large_v2",
+ "whisper_large_v3",
+ "whisper_hindi_large_v2",
+ "whisper_telugu_large_v2",
+ "nemo_english",
+ "nemo_hindi",
+ "vakyansh_bhojpuri",
+ "gcp_v1",
+ "usm",
+ "deepgram",
+ "azure",
+ "seamless_m4t_v2",
+ "mms_1b_all",
+ "seamless_m4t",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/video_bots_page_request_citation_style.py b/src/gooey/types/video_bots_page_request_citation_style.py
new file mode 100644
index 0000000..dc3630b
--- /dev/null
+++ b/src/gooey/types/video_bots_page_request_citation_style.py
@@ -0,0 +1,25 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+VideoBotsPageRequestCitationStyle = typing.Union[
+ typing.Literal[
+ "number",
+ "title",
+ "url",
+ "symbol",
+ "markdown",
+ "html",
+ "slack_mrkdwn",
+ "plaintext",
+ "number_markdown",
+ "number_html",
+ "number_slack_mrkdwn",
+ "number_plaintext",
+ "symbol_markdown",
+ "symbol_html",
+ "symbol_slack_mrkdwn",
+ "symbol_plaintext",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/video_bots_page_request_embedding_model.py b/src/gooey/types/video_bots_page_request_embedding_model.py
new file mode 100644
index 0000000..19c8972
--- /dev/null
+++ b/src/gooey/types/video_bots_page_request_embedding_model.py
@@ -0,0 +1,18 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+VideoBotsPageRequestEmbeddingModel = typing.Union[
+ typing.Literal[
+ "openai_3_large",
+ "openai_3_small",
+ "openai_ada_2",
+ "e5_large_v2",
+ "e5_base_v2",
+ "multilingual_e5_base",
+ "multilingual_e5_large",
+ "gte_large",
+ "gte_base",
+ ],
+ typing.Any,
+]
diff --git a/src/gooey/types/video_bots_page_request_lipsync_model.py b/src/gooey/types/video_bots_page_request_lipsync_model.py
new file mode 100644
index 0000000..3bb98e0
--- /dev/null
+++ b/src/gooey/types/video_bots_page_request_lipsync_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+VideoBotsPageRequestLipsyncModel = typing.Union[typing.Literal["Wav2Lip", "SadTalker"], typing.Any]
diff --git a/src/gooey/types/video_bots_page_request_response_format_type.py b/src/gooey/types/video_bots_page_request_response_format_type.py
new file mode 100644
index 0000000..25cc8f1
--- /dev/null
+++ b/src/gooey/types/video_bots_page_request_response_format_type.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+VideoBotsPageRequestResponseFormatType = typing.Union[typing.Literal["text", "json_object"], typing.Any]
diff --git a/src/gooey/types/video_bots_page_request_translation_model.py b/src/gooey/types/video_bots_page_request_translation_model.py
new file mode 100644
index 0000000..0373c0c
--- /dev/null
+++ b/src/gooey/types/video_bots_page_request_translation_model.py
@@ -0,0 +1,5 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+VideoBotsPageRequestTranslationModel = typing.Union[typing.Literal["google", "ghana_nlp"], typing.Any]
diff --git a/src/gooey/types/video_bots_page_request_tts_provider.py b/src/gooey/types/video_bots_page_request_tts_provider.py
new file mode 100644
index 0000000..3fc8d0a
--- /dev/null
+++ b/src/gooey/types/video_bots_page_request_tts_provider.py
@@ -0,0 +1,7 @@
+# This file was auto-generated by Fern from our API Definition.
+
+import typing
+
+VideoBotsPageRequestTtsProvider = typing.Union[
+ typing.Literal["GOOGLE_TTS", "ELEVEN_LABS", "UBERDUCK", "BARK", "AZURE_TTS", "OPEN_AI"], typing.Any
+]