Swiftpack.co - Package - watson-developer-cloud/swift-sdk

Watson Developer Cloud Swift SDK

Build Status Carthage Compatible Documentation CLA assistant

Overview

The Watson Developer Cloud Swift SDK makes it easy for mobile developers to build Watson-powered applications. With the Swift SDK you can leverage the power of Watson's advanced artificial intelligence, machine learning, and deep learning techniques to understand unstructured data and engage with mobile users in new ways.

There are many resources to help you build your first cognitive application with the Swift SDK:

Contents

General

Services

This SDK provides classes and methods to access the following Watson services.

Before you begin

Requirements

  • Xcode 9.3+
  • Swift 4.1+
  • iOS 10.0+

Installation

The IBM Watson Swift SDK can be installed with Cocoapods, Carthage, or Swift Package Manager.

Cocoapods

You can install Cocoapods with RubyGems:

$ sudo gem install cocoapods

If your project does not yet have a Podfile, use the pod init command in the root directory of your project. To install the Swift SDK using Cocoapods, add the services you will be using to your Podfile as demonstrated below (substituting MyApp with the name of your app). The example below shows all of the currently available services; your Podfile should only include the services that your app will use.

use_frameworks!

target 'MyApp' do
    pod 'IBMWatsonAssistantV1', '~> 0.38.1'
    pod 'IBMWatsonAssistantV2', '~> 0.38.1'
    pod 'IBMWatsonConversationV1', '~> 0.38.1'
    pod 'IBMWatsonDiscoveryV1', '~> 0.38.1'
    pod 'IBMWatsonLanguageTranslatorV3', '~> 0.38.1'
    pod 'IBMWatsonNaturalLanguageClassifierV1', '~> 0.38.1'
    pod 'IBMWatsonNaturalLanguageUnderstandingV1', '~> 0.38.1'
    pod 'IBMWatsonPersonalityInsightsV3', '~> 0.38.1'
    pod 'IBMWatsonSpeechToTextV1', '~> 0.38.1'
    pod 'IBMWatsonTextToSpeechV1', '~> 0.38.1'
    pod 'IBMWatsonToneAnalyzerV3', '~> 0.38.1'
    pod 'IBMWatsonVisualRecognitionV3', '~> 0.38.1'
end

Run the pod install command, and open the generated .xcworkspace file. To update to newer releases, use pod update.

When importing the frameworks in source files, exclude the IBMWatson prefix and the version suffix. For example, after installing IBMWatsonAssistantV1, import it in your source files as import Assistant.

For more information on using Cocoapods, refer to the Cocoapods Guides.

Carthage

You can install Carthage with Homebrew:

$ brew update
$ brew install carthage

If your project does not have a Cartfile yet, use the touch Cartfile command in the root directory of your project. To install the IBM Watson Swift SDK using Carthage, add the following to your Cartfile.

github "watson-developer-cloud/swift-sdk" ~> 0.38.1

Then run the following command to build the dependencies and frameworks:

$ carthage update --platform iOS

Follow the remaining Carthage installation instructions here. Note that the above command will download and build all of the services in the IBM Watson Swift SDK. Make sure to drag-and-drop the built frameworks (only for the services your app requires) into your Xcode project and import them in the source files that require them. The following frameworks need to be added to your app:

  1. RestKit.framework
  2. Whichever services your app will be using (AssistantV1.framework, DiscoveryV1.framework, etc.)
  3. (Speech to Text only) Starscream.framework

Swift Package Manager

Add the following to your Package.swift file to identify the IBM Watson Swift SDK as a dependency. The package manager will clone the Swift SDK when you build your project with swift build.

dependencies: [
    .package(url: "https://github.com/watson-developer-cloud/swift-sdk", from: "0.38.1")
]

Authentication

Watson services are migrating to token-based Identity and Access Management (IAM) authentication.

  • With some service instances, you authenticate to the API by using IAM.
  • In other instances, you authenticate by providing the username and password for the service instance.
  • Visual Recognition uses a form of API key only with instances created before May 23, 2018. Newer instances of Visual Recognition use IAM.

Getting credentials

To find out which authentication to use, view the service credentials. You find the service credentials for authentication the same way for all Watson services:

  1. Go to the IBM Cloud Dashboard page.
  2. Either click an existing Watson service instance or click Create resource > AI and create a service instance.
  3. Click Show to view your service credentials.
  4. Copy the url and either apikey or username and password.

IAM

Some services use token-based Identity and Access Management (IAM) authentication. IAM authentication uses a service API key to get an access token that is passed with the call. Access tokens are valid for approximately one hour and must be regenerated.

You supply either an IAM service API key or an access token:

  • Use the API key to have the SDK manage the lifecycle of the access token. The SDK requests an access token, ensures that the access token is valid, and refreshes it if necessary.
  • Use the access token if you want to manage the lifecycle yourself. For details, see Authenticating with IAM tokens. If you want to switch to API key, override your stored IAM credentials with an IAM API key.

Supplying the IAM API key

let discovery = Discovery(version: "your-version-here", apiKey: "your-apikey-here")

If you are supplying an API key for IBM Cloud Private (ICP), use basic authentication instead, with "apikey" for the username and the api key (prefixed with icp-) for the password. See the Username and Password section.

Supplying the accessToken

let discovery = Discovery(version: "your-version-here", accessToken: "your-accessToken-here")

Updating the accessToken

discovery.accessToken("new-accessToken-here")

Username and Password

let discovery = Discovery(username: "your-username-here", password: "your-password-here", version: "your-version-here")

API Key

Note: This type of authentication only works with Visual Recognition, and for instances created before May 23, 2018. Newer instances of Visual Recognition use IAM.

let visualRecognition = VisualRecognition(apiKey: "your-apiKey-here", version: "your-version-here")

Custom Service URLs

You can set a custom service URL by modifying the serviceURL property. A custom service URL may be required when running an instance in a particular region or connecting through a proxy.

For example, here is how to connect to a Tone Analyzer instance that is hosted in Germany:

let toneAnalyzer = ToneAnalyzer(
    username: "your-username-here",
    password: "your-password-here",
    version: "yyyy-mm-dd"
)
toneAnalyzer.serviceURL = "https://gateway-fra.watsonplatform.net/tone-analyzer/api"

Custom Headers

There are different headers that can be sent to the Watson services. For example, Watson services log requests and their results for the purpose of improving the services, but you can include the X-Watson-Learning-Opt-Out header to opt out of this.

We have exposed a defaultHeaders public property in each class to allow users to easily customize their headers:

let naturalLanguageClassifier = NaturalLanguageClassifier(username: username, password: password)
naturalLanguageClassifier.defaultHeaders = ["X-Watson-Learning-Opt-Out": "true"]

Each service method also accepts an optional headers parameter which is a dictionary of request headers to be sent with the request.

Sample Applications

Synchronous Execution

By default, the SDK executes all networking operations asynchronously. If your application requires synchronous execution, you can use a DispatchGroup. For example:

let dispatchGroup = DispatchGroup()
dispatchGroup.enter()
assistant.message(workspaceID: workspaceID) { response in
    print(response.output.text)
    dispatchGroup.leave()
}
dispatchGroup.wait(timeout: .distantFuture)

Objective-C Compatibility

Please see this tutorial for more information about consuming the Watson Developer Cloud Swift SDK in an Objective-C application.

Linux Compatibility

To use the Watson SDK in your Linux project, please follow the Swift Package Manager instructions.. Note that Speech to Text and Text to Speech are not supported because they rely on frameworks that are unavailable on Linux.

Contributing

We would love any and all help! If you would like to contribute, please read our CONTRIBUTING documentation with information on getting started.

License

This library is licensed under Apache 2.0. Full license text is available in LICENSE.

This SDK is intended for use with an Apple iOS product and intended to be used in conjunction with officially licensed Apple development tools.

Github

link
Stars: 797
Help us keep the lights on

Releases

0.38.1 - Nov 13, 2018

0.38.1 (2018-11-13)

Bug Fixes

  • SpeechToTextV1: Update recognizeMicrophone() to work with any authentication method (5701ba6)

0.38.0 - Nov 12, 2018

0.38.0 (2018-11-12)

Bug Fixes

  • VisualRecognitionV3: temporary workaround for new A12 based devices (62edd09)
  • VisualRecognitionV3: temporary workaround to support A12 devices for Core ML inference (fa212ec)

Features

  • All: Add support for Swift 4.2 (4bbf42b)

0.37.0 - Nov 2, 2018

0.37.0 (2018-11-02)

Features

  • AssistantV1, AssistantV2: Add cloudFunction and webAction to DialogNodeAction.ActionType (16d3fc9)
  • DiscoveryV1: Tokenization dictionaries for collections (d274371)
  • SpeechToTextV1: Add languageCustomizationID parameter to createJob() and recognize() (0137964)

0.36.0 - Oct 19, 2018

0.36.0 (2018-10-19)

Bug Fixes

  • Visual Recognition: Fix deserialization error in getCoreMlModel (9392b23)
  • Conversion of file data to multipart form data (1d46baf)

Features

  • Discovery: Add "LT" option to environment sizes (f92fcde)
  • Discovery: Add size parameter to updateEnvironment method (725e1d5)
  • Discovery: Add bias and loggingOptOut parameters to query methods (8782fc6)
  • Discovery: Add requestedSize and searchStatus properties to Environment model (14cdb02)
  • ToneAnalyzer: Add ToneID options to ToneChatScore (ac75c92)

Reverts

  • AssistantV1: Revert erroneous addition of actions property to OutputData (28efe1e)
  • Linux: Remove 30 minute wait from Linux tests (1b4d734)

0.35.0 - Sep 25, 2018

Adds a new framework for the AssistantV2 service