Microsoft’s Windows 11 AI love-in looks set to continue – here are 3 big risks it needs to avoid

It looks like Microsoft is doubling down on adding artificial intelligence features to Windows 11, with the newly released schedule for Microsoft’s Build 2024 suggesting that many of the talks and presentations will focus on AI and how it can shape the future of computing.

As Windows Central reports, Build 2024, which runs from May 21 to 23, will feature a session called “Designing for a brand new Windows AI feature” which will highlight “brand-new features that allow users deeper interaction with their digital lives on Windows through advance[d] AI features.”

Some publications, such as Windows Latest, suggest that this new AI feature could be the rumored AI Explorer feature. The event will be hosted by Rebecca Del Rio and Adrienne Pauley, both of whom have previously worked on AI projects at Microsoft.

Microsoft’s Build events are primarily aimed at developers and showcase new features and tools that will help them create cutting-edge apps. However, much like Apple’s Worldwide Developers Conference (WWDC), Microsoft Build is often of interest to non-developers as well, as it gives us an insight into what the future holds for Windows 11 (or in WWDC’s case, what’s coming to macOS and iOS).

Judging by the released Build 2024 schedule, it looks like the future of Windows 11 will feature AI – a lot of it. Out of 245 sessions, 79 have the topic of ‘AI Development’ or ‘Copilot’ (Microsoft’s AI-powered assistant), and even sessions such as “Introducing the Next Generation of Windows on Arm” that may not initially seem to be focussed on AI will likely have some references. Arm hardware, which is an alternative to AMD and Intel, is getting improved AI support thanks to NPUs (neural processing units).

The fact that Microsoft is continuing to push AI may not be a surprise to many people – the past few months have seen the company adding all sorts of AI features to various parts of Windows 11, and Build 2024 will be a good chance to prove that it’s still committed to its AI push. However, there are three big risks that Microsoft needs to avoid if it wants to achieve its vision.

The three big risks

broken robot on the floor

(Image credit: Charles Taylor / Shutterstock)

1. Failing to show why AI in Windows 11 is worth it

This is perhaps the biggest risk facing Microsoft. While the company has been adding AI to all parts of Windows 11, I really don’t think the company has shown why I should use these new features.

Copilot, Microsoft’s main AI tool, is now integrated into Windows 11, and is undoubtedly very powerful. However, since Microsoft added it to Windows 11, I’ve used it perhaps two times – once to see how it worked, and a second time to generate an image. Since then, I haven't used it – and it’s not because I distrust AI, as I believe it has enormous and exciting potential.

It’s because Microsoft hasn’t given me a reason to use it. I simply do not know how it could make my day-to-day life easier. Regular tasks I perform on my PC could be improved with some AI help, but I can’t see how right now. 

If Microsoft wants to reap the rewards of all the energy and effort it’s been putting into AI, it needs to show its users why we should be as excited as it is about this brave new world. If it doesn’t, then Copilot, and Windows 11’s other AI tools, could soon be forgotten.

2. Forcing people to use it

Another risk for Microsoft is being too overbearing when it comes to encouraging users to try out the new AI tools. As with my first point, Microsoft must show how AI can improve our lives, not just tell us – and force us to try it out.

Unfortunately, Microsoft has a habit of being rather heavy-handed when it comes to trying to get people to use its software and services. Just look at how it tries to get people to stop downloading the Chrome web browser and use Edge instead, or the increasingly invasive pop-ups in Windows 10 that attempt to convince people to upgrade.

Microsoft has already shown a worrying tendency to do the same with Copilot. Earlier this week we reported on the company changing a fundamental touch gesture in Windows 11 to bring up Copilot, and there are numerous rumors that Microsoft could load Copilot and display it on your desktop when you start Windows 11.

Rather than getting more people to try out AI tools, this aggressive behavior could actually put more people off.

Instead, Microsoft should be confident enough in Copilot to allow users to discover it themselves – while highlighting its virtues without interrupting people while they use Windows 11. And, if a user decides Copilot isn’t for them, Microsoft needs to accept their decision, rather than continuing to nag people in the vain hope that they may change their minds.

3. Losing faith too easily

While Microsoft is going all-in on AI at the moment, we’ve been here before, where Microsoft has identified a new flavor of the month, pushed it on its users, and then given up if it’s not an immediate success.

Just look at Cortana – Microsoft’s previous virtual assistant. It came at a time when its rivals were seeing a lot of success with voice assistants like Apple Siri, Google Assistant, and Amazon Alexa.

Microsoft’s passion meant Cortana was soon tightly integrated into Windows 10 and Windows 11. It had its own taskbar icon and would pop up unbidden when you first set up your new PC. Despite Microsoft’s fervor, it never convinced enough users that Cortana could help improve their lives, and so it ended up being at best an easily forgotten failure, and at worst a useless annoyance. Sound familiar?

Copilot certainly risks falling into the same trap – and the problem is that Microsoft has a reputation for dropping products that fail to take off. Ask any owner of a Microsoft Phone or Zune media player. Or, just look at what ultimately happened to Cortana: the virtual assistant that was once so entwined in Windows that it would pop up as soon as you turned on your new device was slowly hidden away… until it was finally axed and replaced with Copilot.

To avoid this, Microsoft needs to learn from the mistakes it made with Cortana – especially when it comes to convincing its users that Copilot can make a positive impact on their lives. It also needs to have faith that even if Copilot isn’t an immediate hit,  it should continue to invest and improve it, rather than killing it off quickly and moving on to the next thing.

From what I’ve seen and heard from Microsoft, it seems fully committed to AI and Copilot – for now. However, if it doesn’t find a way to prove that AI is the future of Windows, we may see history repeat itself.

You might also like

TechRadar – All the latest technology news

Read More

Google Gemini AI looks like it’s coming to Android tablets and could coexist with Google Assistant (for now)

Google’s new generative AI model, Gemini, is coming to Android tablets. Gemini AI has been observed running on a Google Pixel Tablet, confirming that Gemini can exist on a device alongside Google Assistant… for the time being, at least. Currently, Google Gemini is available to run on Android phones, and it’s expected that it will eventually replace Google Assistant, Google’s current virtual assistant that’s used for voice commands.

When Gemini is installed on Android phones, users would be prompted to choose between using Gemini and Google Assistant. It’s unknown if this restriction will apply to tablets when Gemini finally arrives for them – though at the moment it appears not. 

Man sitting at a table working on a laptop

(Image credit: Shutterstock/GaudiLab)

A discovery in Google Search's code

The news was brought to us via 9to5Google, which did an in-depth report on the latest beta version (15.12) of the Google Search app in the Google Play Store and discovered it contains code referring to using Gemini AI on a “tablet,” and would offer the following features: 

The code also shows that the Google app will host Gemini AI on tablets, instead of a standalone app that currently exists for Android phones. Google might be planning on a separate Gemini app for tablets and possibly other devices, especially if its plans to phase out Google Assistant are still in place. 

9to5Google also warns that this is still as it’s still a beta version of the Google Search app, Google can still change its mind and not roll out these features.

A woman using an Android phone.

(Image credit: Shutterstock/brizmaker)

Where does Google Assistant stand?

When 9to5Google activated Gemini on a Pixel Tablet, it found that Google Assistant and Gemini would function simultaneously. Gemini for Android tablets is yet to be finalized, so Google might implement a similar restriction that prevents both Gemini and Google Assistant running at the same time on tablets. When both were installed and activated, and the voice command “Hey Google” was used, Google Assistant was brought up instead of Gemini.

This in turn contradicted screenshots of the setup screen showing that Gemini will take precedence over Google Assistant if users choose to use it.

The two digital assistants don’t have the same features yet and we know that the Pixel Tablet was designed to act as a smart display that uses Google Assistant when docked. Because Google Assistant will be used when someone asks Gemini to do something it’s unable to do, we may see the two assistants running in parallel for the time being, until Gemini has all of Google Assistant's capabilities, such as smart home features. 

Meanwhile, Android Authority reports that the Gemini experience on the Pixel Tablet is akin to the Pixel Fold and predicts that Google’s tablets will be the first Android to gain Gemini capabilities. This makes sense, as Google may want to use Gemini exclusivity to encourage more people to buy Pixel tablets in the future. The Android tablet market is a highly competitive one, and advanced AI capabilities may help Pixel tablets stand out.

YOU MIGHT ALSO LIKE…

TechRadar – All the latest technology news

Read More

Get ready to learn about what Windows 11 of the future looks like at Microsoft’s March 21 event

We’ve begun getting hints of what Microsoft is gearing up to announce for Windows 11 at its March event, and now we’ve got new pieces of the puzzle. We’re expecting information about a new feature for the Paint app, Paint NPU, and about a feature that’s being referred to as ‘AI Explorer’ internally at Microsoft. 

Microsoft has put up an official page announcing a special digital event named “New Era of Work” which will take place on March 21, starting at 9 PM PDT. On this page, users are met with the tagline “Advancing the new era of work with Copilot” and a description of the event that encourages users to “Tune in here for the latest in scaling AI in your environment with Copilot, Windows, and Surface.”

It sounds like we’re going to get an idea of what the next iteration of Windows Copilot, Microsoft’s new flagship digital AI assistant, will look like and what it’ll be able to do. It also looks like we might see Microsoft’s vision for what AI integration and features will look like for future versions of Windows and Surface products. 

A screenshot of the page announcing Microsoft's digital event.

(Image credit: Microsoft)

What we already know and expect

While we’ll have to wait until the event to see exactly what Microsoft wants to tell us about, we do have some speculation from Windows Latest that one feature we’ll learn about is a Paint app tool powered by new-gen machines’ NPUs (Neural Processing Units). These are processing components that enable new kinds of processes, particularly many AI processes.

This follows earlier reports that indicated that the Paint app was getting an NPU-driven feature, possibly new image editing and rending tools that make use of PCs’ NPUs. Another possible feature that Windows Latest spotted was “LiveCanvas,” which may enable users to draw real-time sketches aided by AI. 

Earlier this week, we also reported about a new ‘AI Explorer’ feature, apparently currently in testing at Microsoft. This new revamped version which has been described as an “advanced Copilot” looks like it could be similar to the Windows Timeline feature, but improved by AI. The present version of Windows Copilot requires an internet connection, but rumors suggest that this could change. 

This is what we currently understand about how the feature will work: it will make records of previous actions users perform, transform them into ‘searchable moments,’ and allow users to search these, as well as retract them. Windows Latest also reinforces the news that most existing PCs running Windows 11 won’t be able to use AI Explorer as it’s designed to use the newest available NPUs, intended to handle and assist higher-level computation tasks. The NPU would enable the AI Explorer feature to work natively on Windows 11 devices and users will be able to interact with AI Explorer using natural language

Using natural language means that users can ask AI Explorer to carry out tasks simply and easily, letting them access past conversations, files, and folders with simple commands, and they will be able to do this with most Windows features and apps. AI Explorer will have the capability to search user history and find information relevant to whatever subject or topic is in the user’s request. We don’t know if it’ll pull this information exclusively from user data or other sources like the internet as well, and we hope this will be clarified on March 21. 

Person working on laptop in kitchen

(Image credit: Getty Images)

What else we might see and what this might mean

 In addition to an NPU-powered Paint app feature and AI Explorer, it looks like we can expect the debut of other AI-powered features including an Automatic Super Resolution feature. This has popped up in Windows 11 23H4 preview builds, and it’s said to leverage PCs’ AI abilities to improve users’ visual experience. This will reportedly be done by utilizing DirectML, an API that also makes use of PCs’ NPUs, and will bring improvements to frame rates in games and apps.

March 21 is gearing up to bring what will at least probably be an exciting presentation, although it’s worth remembering that all of these new features will require an NPU. Only the most newly manufactured Windows devices will come equipped with these, which will leave the overwhelming majority of Windows devices and users in the dust. My guess is Microsoft is really banking on how great the new AI-driven features are to convince users to upgrade to these new models, and with the current state of apps and services like Windows Copilot, that’s still yet to be proven in practice.

YOU MIGHT ALSO LIKE…

TechRadar – All the latest technology news

Read More

Time to start saving – the Apple Vision Pro looks set to launch very soon

It’s time for AR aficionados to start saving, if you haven’t been already, because Apple is getting everything ready to launch the Apple Vision Pro headset in February according to a new report. This follows rumors that it might be delayed until March.

Unnamed sources have told Bloomberg (article behind a paywall) that if things go according to plan, the first wave of consumer units will be ready to ship at the end of January, with a retail release planned for February – sticking to the “early” 2024 release window Apple gave during WWDC 2023 when the headset was unveiled.

It’s apparently not just the product that’s being prepared for an upcoming launch. The report adds that developers creating mixed reality software have recently been told to “get ready” for the Vision Pro, and in January at least two staff members from every US Apple Store branch are supposedly heading to its headquarters for training.

The training is to help them understand the complex Apple gadget. The headset has a lot of customizable components that need to be calibrated and boxed up in-store (online purchases supposedly won't be available) when someone buys one. If there are any problems with the process, potential buyers may walk out the door, or even take home and unbox a subpar experience – something completely unacceptable for a gadget that starts at $ 3,499 per headset (around £2,800 / AU$ 5,300).

But even if the Vision Pro does materialize on shelves in February 2024 you’re unlikely to actually get your hands on one.

A person views an image on a virtual screen while wearing an Apple Vision Pro headset.

Will the Vision Pro replace your TV? (Image credit: Apple)

A February launch for the lucky few

Not simply because the Apple mixed reality headset is priced out of most people’s budgets, but because Apple won’t have many available. 

According to rumors, Apple is only expected to produce 150,000 headsets in 2024. This lack of availability may be why the device will only be sold in the US at launch. And that 150,000 figure is for the whole of 2024; far fewer devices would be available on its release date – so even if you live in the US there’s a good chance you still won’t see one for a while.

The Meta Quest 3 on a notebook surrounded by pens and school supplies on a desk

The Meta Quest 3 is the main Vision Pro rival. (Image credit: Meta)

That is, unless the gadget is wildly unpopular.

While this seems almost impossible for an Apple product, a combination of price and novelty may put people off – even the company’s most rabid fans. What’s more, the headset is certainly the best VR headset ever made from a raw hardware perspective, but Apple has yet to show off software that puts these specs to use in ways that the far cheaper Meta Quest 3 can’t – even its iPhone 15 Pro’s spatial video can play on Quest hardware

It also has some frankly ridiculous problems such as a measly two-hour battery life and (according to some people who have tried it) an uncomfortable design. As I said above, with a $ 3,499 price tag there isn’t any wiggle room – it has to be perfect.

All that said, I’m fully expecting the Apple Vision Pro to be perpetually sold out. This will be Apple’s first new product line in a while, and even if it does wind up being an overpriced folly, Apple collectors will desperately want to get their hands on this piece of tech history.

So if you want to get your hands on one, be ready to book an appointment and head to your local Apple Store as soon as you can. Otherwise, you might have to wait for the Vison Pro's successor to get your hands on an Apple VR headset.

You might also like

TechRadar – All the latest technology news

Read More

Meta’s new VR headset design looks like a next-gen Apple Vision Pro

Meta has teased a super impressive XR headset that looks to combine the Meta Quest Pro, Apple Vision Pro and a few new exclusive features. The only downside? Anything resembling what Meta has shown off is most likely years from release.

During a talk at the University of Arizona College of Optical Sciences, Meta’s director of display systems research, Douglas Lanman, showed a render of Mirror Lake – an advanced prototype that is “practical to build now” based on the tech Meta has developed. This XR headset (XR being a catchall term for VR, AR and MR) combines design elements and features used by the Meta Quest Pro and Apple Vision Pro – such as the Quest Pro’s open side design and the Vision Pro’s EyeSight – with new tools such as HoloCake lenses and electronic varifocal, to make something better than anything on the market.

We’ve talked about electronic varifocal on TechRadar before – when Meta’s Butterscotch Varifocal prototype won an award – so we won’t go too in-depth here. Simply put, using a mixture of eye-tracking and a display system that can move closer or further away from the headset wearer’s face, electronic varifocal aims to mimic the way we focus on objects that are near or far away in the real world. It's an approach Meta calls a “more natural, realistic, and comfortable experience”.

You can see it at work in the video below.

HoloCake lenses help to enable this varifocal system while trimming down the size of the headset – a portmanteau of holographic and pancake.

Pancake lenses are used by the Meta Quest 3, Quest Pro, and other modern headsets including the Pico 4 and Apple Vision Pro, and thanks to some clever optic trickery they can be a lot slimmer than lenses previously used by headsets like the Quest 2.

To further slim the optics down, HoloCake lenses use a thin, flat holographic lens instead of the curved one relied on by a pancake system – holographic as in reflective foil, not as in a 3D hologram you might see in a sci-fi flick.

The only downside is that you need to use lasers, instead of a regular LED backlight. This can add cost, size, heat and safety hurdles. That said, needing to rely on lasers could be seen as an upgrade since these can usually produce a wider and more vivid range of colors than standard LEDs.

A diagram showing the difference between pancake, holocake and regular VR lens optics

Diagrams of different lens optics including HoloCake lenses (Image credit: Meta)

When can we get one? Not for a while 

Unfortunately, Mirror Lake won’t be coming anytime soon. Lanman described the headset as something “[Meta] could build with significant time”, implying that development hasn’t started yet – and even if it has, we might be years away from seeing it in action.

On this point Mark Zuckerberg, Meta’s CEO, added that the technology Mirror Lake relies on could be seen in products “in the second half of the decade”, pointing to a release in 2026 and beyond (maybe late 2025 if we’re lucky).

This would match up with when we predict Meta’s next XR headset – like a Meta Quest Pro or Meta Quest 4 – will probably launch. Meta usually likes to tease its headsets a year in advance at its Meta Connect events (doing so with both the Meta Quest Pro and Quest 3), so if it sticks to this trend the earliest we’ll see a new device is September or October 2025. Meta Connect 2023 passed without a sneak peek at what's to come.

Apple Vision Pro showing a wearer's eye through a display on the front of the headset via EyeSight

Someone wearing the Apple Vision Pro VR headset (Image credit: Apple)

Waiting a few years would also give the Meta Quest 3 time in the spotlight before the next big thing comes to overshadow it, and of course let Meta see how the Apple Vision Pro fares. Apple’s XR headset is taking the exact opposite approach to Meta’s Quest 2 and Quest 3, with Apple offering very high-end tech at a very unaffordable price ($ 3,499, or around £2,800 / AU$ 5,300). 

If Apple’s gamble pays off, Meta might want to mix up its strategy by releasing an equally high-end and costly Meta Quest Pro 2 that offers a more significant upgrade over the Quest 3 than the first Meta Quest Pro offered compared to the Quest 2. If the Vision Pro flops, Meta won’t want to follow its lead.

YOU MIGHT ALSO LIKE

TechRadar – All the latest technology news

Read More

WhatsApp’s built-in AI chatbot looks like its rolling out to more people

AI bots are rapidly being added to just about every app and platform you can think of – with more on the way – and WhatsApp is stepping up its testing of a chatbot of its own, with easier access to the feature now on the way.

Back in September, WhatsApp owner Meta announced a variety of AI upgrades that would be coming to its products. Since then, a small number of users have been able to play around with an AI chatbot inside WhatsApp, capable of answering questions, generating text, and creating art like stickers.

Now, as spotted by WABetaInfo (via Android Police), a shortcut to the AI chat functionality has been added to the conversations screen in the beta version of WhatsApp for Android. If you're running the early beta version of the app, you may see it soon.

It also means that it shouldn't be too long before the rest of us get the same feature, and we can see how WhatsApp's AI helper compares against the likes of ChatGPT and Google Bard when it comes to providing useful and accurate information.

WhatsApp and AI

From what Meta has said so far, the purpose of the AI chatbot inside WhatsApp is to help with daily activities, offering advice and suggestions: how to entertain the kids at the weekend perhaps, or what to look for when upgrading a smartphone.

WhatsApp is by no means the first messaging app to give this a try – Snapchat introduced a similar feature back in February, and the chats with the AI buddy appears alongside the rest of your conversations through the app.

Such are the capabilities of generative AI now, you can really ask these bots anything you like – from relationship advice to questions about complex technical topics. The point of them being built into apps is that you're less likely to leave the app and go somewhere else to get your AI-produced responses.

WhatsApp continues to be one of the most regularly updated apps out there: we've recently seen AI-made chat stickers, newsletter tools, and features to fight spammers introduced for users of the instant messenger.

Join TechRadar’s WhatsApp Channel to get the latest tech news, reviews, opinion and Black Friday deals directly from our editors. We’ll make sure you don’t miss a thing!

You might also like

TechRadar – All the latest technology news

Read More

Google Maps now looks more like Apple Maps – and a lot of people aren’t happy

Google Maps has had a rejig of the colors used to denote different elements, and a significant portion of its regular users aren't happy about the change.

As you may have seen, this change in color palette was first spotted back in September, but now it’s widely rolling out to users of Google’s navigation app.

Google Maps now has gray roads like Apple, rather than white or yellow roads as before, and forests are a darker green. On the other hand, the shade of blue used for water is lighter.

However, the active route is a much darker blue, with alternate routes shown in lighter blue (these used to be gray).

See the pic above for a comparison of the old (left) and new (right) design, and the one below (in the tweet) for another look at the freshly revamped colors.

These may not sound like massive changes – and to be fair, they aren’t, they’re essentially tweaks. But they have rubbed a number of users up the wrong way. As Android Authority points out, there’s some quite spicy feedback on the new Google Maps on Reddit, X (formerly Twitter) and other online forums.


See more

Analysis: Lacking clarity?

Some of this is unfamiliarity, as no one likes change, and it takes time to acclimatize to a new look – but there are some consistent and well-observed pieces of feedback on the redeployment of colors for Google Maps.

One common thread is criticism of the new colors lacking clarity, and making it trickier to see what’s what at a glance (and when driving obviously you will just be glancing at the display).

As one Redditor put it: “I’m finding it a little hard to read as quickly as I used to. The toned down look is cute but not practical.”

Another problem highlighted by multiple users on Reddit is that the new alternate routes being blue – as well as the main route, albeit that’s a darker blue – is an issue. It can be difficult to tell those routes apart on a phone at a bit of a distance (and with other potential factors thrown into the mix like sun glare).

Overall, Google may want to have a rethink, particularly around the alternate routes. That said, not everyone is unhappy with the changes, but the majority seem to be at least according to a poll Android Authority is running.

This shows that 44% of respondents don’t like the new colors, compared to 28% who do (with the rest abstaining). So, that doesn’t look great for Google, though of course, it’s a limited sample of around 800 people (at the time of writing).

You might also like

TechRadar – All the latest technology news

Read More

Adobe’s new photo editor looks even more powerful than Google’s Magic Editor

Adobe MAX 2023 is less than a week away, and to promote the event, the company recently published a video teasing its new “object-aware editing engine” called Project Stardust.

According to the trailer, the feature has the ability to identify individual objects in a photograph and instantly separate them into their own layers. Those same objects can then be moved around on-screen or deleted. Selecting can be done either manually or automatically via the Remove Distractions tool. The software appears to understand the difference between the main subjects in an image and the people in the background that you want to get rid of.

What’s interesting is moving or deleting something doesn’t leave behind a hole. The empty space is filled in most likely by a generative AI model. Plus, you can clean up any left-behind evidence of a deleted item. In its sample image, Adobe erases a suitcase held by a female model and then proceeds to edit her hand so that she’s holding a bouquet of flowers instead.  

Image 1 of 2

Project Stardust editing

(Image credit: Adobe)
Image 2 of 2

Project Stardust generative AI

(Image credit: Adobe)

The same tech can also be used to change articles of clothing in pictures. A yellow down jacket can be turned into a black leather jacket or a pair of khakis into black jeans. To do this, users will have to highlight the piece of clothing and then enter what they want to see into a text prompt. 

Stardust replacement tool

(Image credit: Adobe)

AI editor

Functionally, Project Stardust operates similarly to Google’s Magic Editor which is a generative AI tool present on the Pixel 8 series. The tool lets users highlight objects in a photograph and reposition them in whatever manner they please. It, too, can fill gaps in images by creating new pixels. However, Stardust feels much more capable. The Pixel 8 Pro’s Magic Eraser can fill in gaps, but neither it nor Magic Editor can’t generate content. Additionally, Google’s version requires manual input whereas Adobe’s software doesn’t need it.

Seeing these two side-by-side, we can’t but wonder if Stardust is actually powered by Google’s AI tech. Very recently, the two companies announced they were entering a partnership “and offering a free three-month trial for Photoshop on the web for people who buy a Chromebook Plus device. Perhaps this “partnership” runs a lot deeper than free Photoshop considering how similar Stardust is to Magic Editor.

Impending reveal

We should mention that Stardust isn't perfect. If you look at the trailer, you'll notice some errors like random holes in the leather jacket and strange warping around the flower model's hands. But maybe what we see is Stardust in an early stage. 

There is still a lot we don’t know like whether it's a standalone app or will it be housed in, say, Photoshop? Is Stardust releasing in beta first or are we getting the final version? All will presumably be answered on October 10 when Adobe MAX 2023 kicks off. What’s more, the company will be showing other “AI features” coming to “Firefly, Creative Cloud, Express, and more.”

Be sure to check out TechRadar’s list of the best Photoshop courses online for 2023 if you’re thinking of learning the software, but don’t know where to start. 

You might also like

TechRadar – All the latest technology news

Read More

Procreate is releasing a new iPad app for animation – and it looks incredible

Procreate, the best-selling paid app on iPad for over six years, has unveiled an exciting addition to its arsenal – a new animation app called Procreate Dreams.

Following in the footsteps of the widely acclaimed Procreate app, which is a more accessible alternative to traditional digital art software that still offers powerful and complex tools, Procreate Dreams will deliver a touch-centric creative experience.

The app is slated for release on November 22, 2023, and much like its two-time Apple Design Award-winning predecessor, it will be available for a one-time purchase of $ 19.99 (UK and AU prices TBC). In stark contrast to more complex Adobe software, there are no costly subscriptions with Procreate

The best iPads offer incredible computing power, great pressure sensitivity, and excellent accuracy, increasingly making them a go-to tool for artists at all skill levels. Procreate Dreams will similarly leverage Apple's slate – and in particular, those featuring Apple Silicon chips – to deliver smooth and efficient workflows as well as intuitive tools suitable for all skill levels.

Procreate dream keynote

(Image credit: Future)

New features and tools

To create a more accessible animation experience, Procreate has created several features and tools for Dreams, some of which may be familiar to users of the original app.

One such tool is Performing, which allows creatives to animate through touch, automatically adding keyframes while recording actions. This breathes life into the artwork in real-time, and without the need for complex keyframes and paths, making it a lot more approachable for inexperienced animators.

Also noteworthy is the app's Multi-touch Timeline, which allows for quick and easy navigation, organization, and editing. Procreate claims this system is faster than a mouse and keyboard, and that for the first time, it allows artists and animators to combine drawing, cel animation, keyframing, video editing, and compositing, all performed with fluid gestures.

By drawing directly on the timeline with the Apple Pencil, you can quickly select content, tracks, or keyframes, and then edit them all at once. Group drawings, move tracks, retime content, adjust keyframes, and much more. The timeline is kept clutter-free with simplified keyframes, but you can also tap into a keyframe for more precise property edits.

Its painting and compositing engine is immensely powerful, now allowing creatives to work on raster projects with resolutions of up to 1 million x 1 million pixels with abundant layers. It also supports all of Procreate's brushes, which are Apple Pencil-optimized. There will also be full support for files created in Procreate.

Plus, thanks to GPU acceleration built on Metal and powered by Apple Silicon, creatives will be able to instantly play back projects as they edit. This level of real-time rendering stands to save animators a lot of time wasted on RAM previews.
Procreate Dreams introduces a new audio engine, enabling voice-overs, atmospheric music, and sound effects to bring creations to life. High-resolution video editing is also seamless with ProRes footage up to 8K supported. Video can be drawn over, allowing you to create annotated or rotoscoped projects using the full breadth of Procreate's brush library,

There's plenty more to shout about, too; Flipbook, which pays homage to traditional animation and builds on the original Animation Assist tool in Procreate; text and typography; and an iCloud synchronized Procreate file format, which the company says sees 1TB files opened in an instant. At long last, this new file format will offer eternal undo history, too.

With Procreate Dreams, the boundaries of what's possible in animation and digital art might be about to change beyond belief, much like what Procreate's original app did for digital art – so keep your eyes peeled for our first impressions of the app when it releases later this year!

Procreate dream keynote

(Image credit: Future)

Analysis: Procreate joins the big leagues

It's been 10 years since Procreate's first app was released, and in that time it's gone from strength to strength. Given it featured an animation assist feature, it always felt like just a matter of time until the developer released a full animation app – so why now?

Well, the unrivaled computing power of the Apple M2 chips – especially in the tablet market – is likely the culprit here. The best graphics tablets used for animation will generally need to be hooked up to a powerful computer that handles the processing and rendering of beefy animation software, as the slate itself is just the canvas and workflow interface. Of course, with iPads, that's not the case; these lightweight slates handle everything on board.

Animation, especially at the scale Procreate Dreams is offering, requires immense computing power, and up until now it likely just wasn't possible; certainly not in a way that fully optimized the iPad's innate features and benefits while also remaining accessible to non-professionals, which is a key selling point for Procreate's software.

Given how long this has inevitably been in the pipeline, and just how successful the original app was as a gateway to digital art for so many – including myself – it's hard to imagine the software being a flop. We'll see come November.

TechRadar – All the latest technology news

Read More

Samsung’s leaked XR headset looks like a very budget Apple Vision Pro

At Galaxy Unpacked back in February 2023 Samsung announced that it was partnering with Google and Qualcomm to develop a XR headset – and if these leaks are correct we might have got our first look at its efforts so far. Well, kinda.

Very little is know about Samsung’s latest XR efforts, though it’s expected to be a standalone device rather than something powered by smartphones like its old Gear VR headsets. This new leaked prototype (assuming it’s real) gives us an insight into Samsung’s design philosophy, though according to the leaks it’s not an in-development prototype. Instead, this is what Samsung was working on until it saw the Apple Vision Pro and decided to start over.

The leak was first posted by the Chinese publication Vrtuoluo, but the article has been deleted (via Android Authority). The only way to view the original is using the Way Back Machine which has archived the original. The images it posted look a lot like a typical VR headset – such as the Oculus Quest 2 – with it apparently featuring four tracking cameras, dual RGB cameras, and a depth sensor for full-color passthrough. No controllers are included with the images so it appears that much like the Vision Pro the Samsung XR device would use hand and eye-tracking controls by default.

A VR headset cla in black plastic with a simple strap and six visible cameras on its faces

(Image credit: Vrtuoluo / Samsung)

The prototype apparently also uses dual micro OLED displays, and (surprisingly) a Samsung Exynos 2200 rather than a Qualcomm XR chip such as the Snapdragon XR 2 found in many VR headsets like the Pico 4, or a XR 2 Plus like the one in the Meta Quest Pro

It’s worth remembering this is just a leak however, and not one that we may ever be able to easily verify – as this is a prototype for a headset that we should never see publicly. As such we should take the information and images with a pinch of salt. That said, if this is indeed a canceled Samsung XR headset, we can see why the project is no longer in development.

Not an Apple Vision Pro rival yet

This leaked Samsung prototype isn’t close to being a Vision Pro competitor. The specs are fine, but not in the same league as the Apple headset, and the design is significantly more bulky. After seeing the Vision Pro announcement we can see why Samsung might want to go back to the drawing board.

Not being a Vision Pro rival isn’t necessarily a bad thing, however. Sure, borrowing some of its cues could be smart, but mimic it too heavily and you’ll end up copying its biggest flaw – the sky-high price of $ 3,499 (around £2,800 / AU$ 5,300).

If a Samsung headset can offer many of the Vision Pro’s features at a fraction of its cost – with rumors teasing it might cost closer to $ 1,000 / £1,000 / AU$ 1,500 – then it could be onto a winner. It’ll likely be a while before we see anything from Samsung though. With it abandoning this nearly complete project for a new one it’s possible we won’t know anything concrete until 2024 or even later – we’ll just have to wait and see what it announces. 

TechRadar – All the latest technology news

Read More