Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
127 changes: 127 additions & 0 deletions genai/live/live_conversation_audio_with_audio.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,127 @@
# Copyright 2025 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

# [START googlegenaisdk_live_conversation_audio_with_audio]

import asyncio
import base64

from google import genai
from google.genai.types import (AudioTranscriptionConfig, Blob,
LiveConnectConfig, Modality)
import numpy as np

from scipy.io import wavfile

# The number of audio frames to send in each chunk.
CHUNK = 4200
CHANNELS = 1
MODEL = "gemini-2.0-flash-live-preview-04-09"
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Update! Can you use gemini-2.5-flash-image-preview


# The audio sample rate expected by the model.
INPUT_RATE = 16000
# The audio sample rate of the audio generated by the model.
OUTPUT_RATE = 24000

# The sample width for 16-bit audio, which is standard for this type of audio data.
SAMPLE_WIDTH = 2

client = genai.Client()


def read_wavefile(filepath: str) -> tuple[str, str]:
# Read the .wav file using scipy.io.wavfile.read
rate, data = wavfile.read(filepath)
# Convert the NumPy array of audio samples back to raw bytes
raw_audio_bytes = data.tobytes()
# Encode the raw bytes to a base64 string.
# The result needs to be decoded from bytes to a UTF-8 string
base64_encoded_data = base64.b64encode(raw_audio_bytes).decode("ascii")
mime_type = f"audio/pcm;rate={rate}"
return base64_encoded_data, mime_type


def write_wavefile(filepath: str, audio_frames: list[bytes], rate: int) -> None:
"""Writes a list of audio byte frames to a WAV file using scipy."""
# Combine the list of byte frames into a single byte string
raw_audio_bytes = b"".join(audio_frames)

# Convert the raw bytes to a NumPy array.
# The sample width is 2 bytes (16-bit), so we use np.int16
audio_data = np.frombuffer(raw_audio_bytes, dtype=np.int16)

# Write the NumPy array to a .wav file
wavfile.write(filepath, rate, audio_data)
print(f"Model response saved to {filepath}")


async def main() -> bool:
print("Starting the code")
async with client.aio.live.connect(
model=MODEL,
config=LiveConnectConfig(
# Set Model responses to be in Audio
response_modalities=[Modality.AUDIO],
# To generate transcript for input audio
input_audio_transcription=AudioTranscriptionConfig(),
# To generate transcript for output audio
output_audio_transcription=AudioTranscriptionConfig(),
),
) as session:

async def send() -> None:
# using local file as an example for live audio input
wav_file_path = "hello_gemini_are_you_there.wav"
base64_data, mime_type = read_wavefile(wav_file_path)
audio_bytes = base64.b64decode(base64_data)
await session.send_realtime_input(media=Blob(data=audio_bytes, mime_type=mime_type))

async def receive() -> None:
audio_frames = []

async for message in session.receive():
if message.server_content.input_transcription:
print(message.server_content.model_dump(mode="json", exclude_none=True))
if message.server_content.output_transcription:
print(message.server_content.model_dump(mode="json", exclude_none=True))
if message.server_content.model_turn:
for part in message.server_content.model_turn.parts:
if part.inline_data.data:
audio_data = part.inline_data.data
audio_frames.append(audio_data)

if audio_frames:
write_wavefile(
"example_model_response.wav",
audio_frames,
OUTPUT_RATE,
)

send_task = asyncio.create_task(send())
receive_task = asyncio.create_task(receive())
await asyncio.gather(send_task, receive_task)
# Example response:
# gemini-2.0-flash-live-preview-04-09
# {'input_transcription': {'text': 'Hello.'}}
# {'output_transcription': {}}
# {'output_transcription': {'text': 'Hi'}}
# {'output_transcription': {'text': ' there. What can I do for you today?'}}
# {'output_transcription': {'finished': True}}
# Model response saved to example_model_response.wav

# [END googlegenaisdk_live_conversation_audio_with_audio]
return True

if __name__ == "__main__":
asyncio.run(main())
66 changes: 66 additions & 0 deletions genai/live/live_ground_ragengine_with_txt.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,66 @@
# Copyright 2025 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import asyncio

_memory_corpus = "projects/cloud-ai-devrel-softserve/locations/us-central1/ragCorpora/2305843009213693952"


async def generate_content(memory_corpus: str) -> list[str]:
# [START googlegenaisdk_live_ground_ragengine_with_txt]
from google import genai
from google.genai.types import (Content, LiveConnectConfig, Modality, Part,
Retrieval, Tool, VertexRagStore,
VertexRagStoreRagResource)

client = genai.Client()
model_id = "gemini-2.0-flash-live-preview-04-09"
rag_store = VertexRagStore(
rag_resources=[
VertexRagStoreRagResource(
rag_corpus=memory_corpus # Use memory corpus if you want to store context.
)
],
# Set `store_context` to true to allow Live API sink context into your memory corpus.
store_context=True,
)
config = LiveConnectConfig(
response_modalities=[Modality.TEXT],
tools=[Tool(retrieval=Retrieval(vertex_rag_store=rag_store))],
)

async with client.aio.live.connect(model=model_id, config=config) as session:
text_input = "What year did Mariusz Pudzianowski win World's Strongest Man?"
print("> ", text_input, "\n")

await session.send_client_content(
turns=Content(role="user", parts=[Part(text=text_input)])
)

response = []

async for message in session.receive():
if message.text:
response.append(message.text)
continue

print("".join(response))
# Example output:
# > What year did Mariusz Pudzianowski win World's Strongest Man?
# Mariusz Pudzianowski won World's Strongest Man in 2002, 2003, 2005, 2007, and 2008.
# [END googlegenaisdk_live_ground_ragengine_with_txt]
return response


if __name__ == "__main__":
asyncio.run(generate_content(_memory_corpus))
16 changes: 8 additions & 8 deletions genai/live/live_websocket_audiogen_with_txt.py
Original file line number Diff line number Diff line change
Expand Up @@ -55,9 +55,7 @@ async def generate_content() -> str:

# Websocket Configuration
WEBSOCKET_HOST = "us-central1-aiplatform.googleapis.com"
WEBSOCKET_SERVICE_URL = (
f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"
)
WEBSOCKET_SERVICE_URL = f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"

# Websocket Authentication
headers = {
Expand All @@ -66,9 +64,7 @@ async def generate_content() -> str:
}

# Model Configuration
model_path = (
f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
)
model_path = f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
model_generation_config = {
"response_modalities": ["AUDIO"],
"speech_config": {
Expand Down Expand Up @@ -129,15 +125,19 @@ async def generate_content() -> str:
for part in model_turn["parts"]:
if part["inlineData"]["mimeType"] == "audio/pcm":
audio_chunk = base64.b64decode(part["inlineData"]["data"])
aggregated_response_parts.append(np.frombuffer(audio_chunk, dtype=np.int16))
aggregated_response_parts.append(
np.frombuffer(audio_chunk, dtype=np.int16)
)

# End of response
if server_content.get("turnComplete"):
break

# Save audio to a file
if aggregated_response_parts:
wavfile.write("output.wav", 24000, np.concatenate(aggregated_response_parts))
wavfile.write(
"output.wav", 24000, np.concatenate(aggregated_response_parts)
)
# Example response:
# Setup Response: {'setupComplete': {}}
# Input: Hello? Gemini are you there?
Expand Down
12 changes: 5 additions & 7 deletions genai/live/live_websocket_audiotranscript_with_txt.py
Original file line number Diff line number Diff line change
Expand Up @@ -55,9 +55,7 @@ async def generate_content() -> str:

# Websocket Configuration
WEBSOCKET_HOST = "us-central1-aiplatform.googleapis.com"
WEBSOCKET_SERVICE_URL = (
f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"
)
WEBSOCKET_SERVICE_URL = f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"

# Websocket Authentication
headers = {
Expand All @@ -66,9 +64,7 @@ async def generate_content() -> str:
}

# Model Configuration
model_path = (
f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
)
model_path = f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
model_generation_config = {
"response_modalities": ["AUDIO"],
"speech_config": {
Expand Down Expand Up @@ -142,7 +138,9 @@ async def generate_content() -> str:
for part in model_turn["parts"]:
if part["inlineData"]["mimeType"] == "audio/pcm":
audio_chunk = base64.b64decode(part["inlineData"]["data"])
aggregated_response_parts.append(np.frombuffer(audio_chunk, dtype=np.int16))
aggregated_response_parts.append(
np.frombuffer(audio_chunk, dtype=np.int16)
)

# End of response
if server_content.get("turnComplete"):
Expand Down
12 changes: 5 additions & 7 deletions genai/live/live_websocket_textgen_with_audio.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,9 +65,7 @@ def read_wavefile(filepath: str) -> tuple[str, str]:

# Websocket Configuration
WEBSOCKET_HOST = "us-central1-aiplatform.googleapis.com"
WEBSOCKET_SERVICE_URL = (
f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"
)
WEBSOCKET_SERVICE_URL = f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"

# Websocket Authentication
headers = {
Expand All @@ -76,9 +74,7 @@ def read_wavefile(filepath: str) -> tuple[str, str]:
}

# Model Configuration
model_path = (
f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
)
model_path = f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
model_generation_config = {"response_modalities": ["TEXT"]}

async with connect(WEBSOCKET_SERVICE_URL, additional_headers=headers) as websocket_session:
Expand All @@ -105,7 +101,9 @@ def read_wavefile(filepath: str) -> tuple[str, str]:
return "Error: WebSocket setup failed."

# 3. Send audio message
encoded_audio_message, mime_type = read_wavefile("hello_gemini_are_you_there.wav")
encoded_audio_message, mime_type = read_wavefile(
"hello_gemini_are_you_there.wav"
)
# Example audio message: "Hello? Gemini are you there?"

user_message = {
Expand Down
8 changes: 2 additions & 6 deletions genai/live/live_websocket_textgen_with_txt.py
Original file line number Diff line number Diff line change
Expand Up @@ -52,9 +52,7 @@ async def generate_content() -> str:

# Websocket Configuration
WEBSOCKET_HOST = "us-central1-aiplatform.googleapis.com"
WEBSOCKET_SERVICE_URL = (
f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"
)
WEBSOCKET_SERVICE_URL = f"wss://{WEBSOCKET_HOST}/ws/google.cloud.aiplatform.v1.LlmBidiService/BidiGenerateContent"

# Websocket Authentication
headers = {
Expand All @@ -63,9 +61,7 @@ async def generate_content() -> str:
}

# Model Configuration
model_path = (
f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
)
model_path = f"projects/{PROJECT_ID}/locations/{LOCATION}/publishers/google/models/{GEMINI_MODEL_NAME}"
model_generation_config = {"response_modalities": ["TEXT"]}

async with connect(WEBSOCKET_SERVICE_URL, additional_headers=headers) as websocket_session:
Expand Down
1 change: 1 addition & 0 deletions genai/live/requirements-test.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,3 +2,4 @@ backoff==2.2.1
google-api-core==2.25.1
pytest==8.4.1
pytest-asyncio==1.1.0
pytest-mock==3.14.0
15 changes: 8 additions & 7 deletions genai/live/requirements.txt
Original file line number Diff line number Diff line change
@@ -1,7 +1,8 @@
google-genai==1.28.0
scipy==1.16.1
websockets==15.0.1
numpy==1.26.4
soundfile==0.12.1
openai==1.99.1
setuptools==80.9.0
google-genai==1.28.0
scipy==1.16.1
websockets==15.0.1
numpy==1.26.4
soundfile==0.12.1
openai==1.99.1
setuptools==80.9.0
pyaudio==0.2.14
Loading