| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181 |
- // Copyright 2023 Google LLC
- //
- // Licensed under the Apache License, Version 2.0 (the "License");
- // you may not use this file except in compliance with the License.
- // You may obtain a copy of the License at
- //
- // http://www.apache.org/licenses/LICENSE-2.0
- //
- // Unless required by applicable law or agreed to in writing, software
- // distributed under the License is distributed on an "AS IS" BASIS,
- // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- // See the License for the specific language governing permissions and
- // limitations under the License.
- import Foundation
- /// An object that represents a back-and-forth chat with a model, capturing the history and saving
- /// the context in memory between each message sent.
- @available(iOS 15.0, macOS 12.0, macCatalyst 15.0, tvOS 15.0, watchOS 8.0, *)
- public final class Chat: Sendable {
- private let model: GenerativeModel
- /// Initializes a new chat representing a 1:1 conversation between model and user.
- init(model: GenerativeModel, history: [ModelContent]) {
- self.model = model
- self.history = history
- }
- private let historyLock = NSLock()
- private nonisolated(unsafe) var _history: [ModelContent] = []
- /// The previous content from the chat that has been successfully sent and received from the
- /// model. This will be provided to the model for each message sent as context for the discussion.
- public var history: [ModelContent] {
- get {
- historyLock.withLock { _history }
- }
- set {
- historyLock.withLock { _history = newValue }
- }
- }
- private func appendHistory(contentsOf: [ModelContent]) {
- historyLock.withLock {
- _history.append(contentsOf: contentsOf)
- }
- }
- /// Sends a message using the existing history of this chat as context. If successful, the message
- /// and response will be added to the history. If unsuccessful, history will remain unchanged.
- /// - Parameter parts: The new content to send as a single chat message.
- /// - Returns: The model's response if no error occurred.
- /// - Throws: A ``GenerateContentError`` if an error occurred.
- public func sendMessage(_ parts: any PartsRepresentable...) async throws
- -> GenerateContentResponse {
- return try await sendMessage([ModelContent(parts: parts)])
- }
- /// Sends a message using the existing history of this chat as context. If successful, the message
- /// and response will be added to the history. If unsuccessful, history will remain unchanged.
- /// - Parameter content: The new content to send as a single chat message.
- /// - Returns: The model's response if no error occurred.
- /// - Throws: A ``GenerateContentError`` if an error occurred.
- public func sendMessage(_ content: [ModelContent]) async throws
- -> GenerateContentResponse {
- // Ensure that the new content has the role set.
- let newContent = content.map(populateContentRole(_:))
- // Send the history alongside the new message as context.
- let request = history + newContent
- let result = try await model.generateContent(request)
- guard let reply = result.candidates.first?.content else {
- let error = NSError(domain: "com.google.generative-ai",
- code: -1,
- userInfo: [
- NSLocalizedDescriptionKey: "No candidates with content available.",
- ])
- throw GenerateContentError.internalError(underlying: error)
- }
- // Make sure we inject the role into the content received.
- let toAdd = ModelContent(role: "model", parts: reply.parts)
- // Append the request and successful result to history, then return the value.
- appendHistory(contentsOf: newContent)
- history.append(toAdd)
- return result
- }
- /// Sends a message using the existing history of this chat as context. If successful, the message
- /// and response will be added to the history. If unsuccessful, history will remain unchanged.
- /// - Parameter parts: The new content to send as a single chat message.
- /// - Returns: A stream containing the model's response or an error if an error occurred.
- @available(macOS 12.0, *)
- public func sendMessageStream(_ parts: any PartsRepresentable...) throws
- -> AsyncThrowingStream<GenerateContentResponse, Error> {
- return try sendMessageStream([ModelContent(parts: parts)])
- }
- /// Sends a message using the existing history of this chat as context. If successful, the message
- /// and response will be added to the history. If unsuccessful, history will remain unchanged.
- /// - Parameter content: The new content to send as a single chat message.
- /// - Returns: A stream containing the model's response or an error if an error occurred.
- @available(macOS 12.0, *)
- public func sendMessageStream(_ content: [ModelContent]) throws
- -> AsyncThrowingStream<GenerateContentResponse, Error> {
- // Ensure that the new content has the role set.
- let newContent: [ModelContent] = content.map(populateContentRole(_:))
- // Send the history alongside the new message as context.
- let request = history + newContent
- let stream = try model.generateContentStream(request)
- return AsyncThrowingStream { continuation in
- Task {
- var aggregatedContent: [ModelContent] = []
- do {
- for try await chunk in stream {
- // Capture any content that's streaming. This should be populated if there's no error.
- if let chunkContent = chunk.candidates.first?.content {
- aggregatedContent.append(chunkContent)
- }
- // Pass along the chunk.
- continuation.yield(chunk)
- }
- } catch {
- // Rethrow the error that the underlying stream threw. Don't add anything to history.
- continuation.finish(throwing: error)
- return
- }
- // Save the request.
- appendHistory(contentsOf: newContent)
- // Aggregate the content to add it to the history before we finish.
- let aggregated = self.aggregatedChunks(aggregatedContent)
- self.history.append(aggregated)
- continuation.finish()
- }
- }
- }
- private func aggregatedChunks(_ chunks: [ModelContent]) -> ModelContent {
- var parts: [any Part] = []
- var combinedText = ""
- for aggregate in chunks {
- // Loop through all the parts, aggregating the text and adding the images.
- for part in aggregate.parts {
- switch part {
- case let textPart as TextPart:
- combinedText += textPart.text
- default:
- // Don't combine it, just add to the content. If there's any text pending, add that as
- // a part.
- if !combinedText.isEmpty {
- parts.append(TextPart(combinedText))
- combinedText = ""
- }
- parts.append(part)
- }
- }
- }
- if !combinedText.isEmpty {
- parts.append(TextPart(combinedText))
- }
- return ModelContent(role: "model", parts: parts)
- }
- /// Populates the `role` field with `user` if it doesn't exist. Required in chat sessions.
- private func populateContentRole(_ content: ModelContent) -> ModelContent {
- if content.role != nil {
- return content
- } else {
- return ModelContent(role: "user", parts: content.parts)
- }
- }
- }
|