New iPhone feature can clone your voice in just 15 minutes

In this article:
Other accessibility features include Assistive Access, Live Speech and more   (Apple )
Other accessibility features include Assistive Access, Live Speech and more (Apple )

Apple has unveiled a new iOS feature that allows iPhone, iPad and Mac users to make a digital clone of their voice in just 15 minutes.

The Personal Voice feature announced on Tuesday is aimed at helping those who are at risk of losing their voice, such as people with an amyotrophic lateral sclerosis (ALS) diagnosis or other neurological conditions.

The new feature forms part of a suite of updates announced ahead of Global Accessibility Awareness Day on 18 May, and will likely be rolled out as part of iOS 17 later this year.

The Motor Neurone Disease Association suggests that cloning a voice, known as voice banking, typically takes an average of two hours or longer with current tech, and can cost hundreds of pounds.

The Independent got a first look at the new feature in action, and the Personal Voice sounded remarkably similar to the user’s actual voice, albeit with a slightly robotic, synthesised tone.

It can be trained in 15 minutes simply by reading a bunch of randomly generated text prompts, and uses on-device machine learning to ensure your data is kept private and secure.

Personal Voice will be available for use with Apple’s new Live Speech feature on iOS, which will let users type what they want to say and have it spoken out loud, during in-person conversations, as well as with phone and FaceTime calls. The feature is said to work with all accents and dialects.

 (Apple)
(Apple)

As well as features for speech accessibility, Apple announced a cognitive disability setting for the iPhone and iPad called Assistive Access, which essentially lets you pare your device down to a few core apps of your choosing.

Arranged in a grid or a list, it simplifies the user interface down to large, easily readable buttons, so you can make calls, access an emoji-only keyboard, and a fuss-free camera, without the clutter.

Companies such as Doro and Jitterbug currently have a stranglehold on the “simple phone”, offering devices for senior users that strip away the bulk, and deliver a simple user interface with large high-contrast buttons that make it easy to use a phone.

With Apple’s new Assistive Access feature, those with cognitive disabilities will be able to take advantage of a mainstream iPhone device and its features without having to opt for an Android device or one designed for their specific need.

 (Apple)
(Apple)

In addition, Apple previewed a new feature in the Magnifier app for low vision users called Point and Speak, which makes use of the camera, the LiDAR Scanner and on-device machine learning to read aloud text that a user places their finger on.

Plus, the company announced Mac support for Made for iPhone hearing devices, phonetic suggestions for Voice Control users and the ability for Switch Control users to turn their switches into game controllers on the iPhone and iPad.

Apple has made it a tradition over the years to unveil new accessibility features ahead of WWDC in June, where it usually reveals the next iOS update, emphasising its prioritisation of accessibility within the iOS ecosystem.

While  Apple didn’t state when exactly the new updates would be coming to its devices later this year, a rollout alongside iOS 17 seems likely, given previous announcements.

Advertisement