Microsoft could be planning to infuse Windows 11 with AI in every corner

Windows 11 could get AI enhancements for a bunch of its core apps in the future, according to a well-known source of leaks on all things Microsoft.

Zac Bowden (of Windows Central) informs us Microsoft is considering bolstering a ‘handful’ of the default apps in Windows 11 with AI functionality, and that includes Photos, Snipping Tool, Paint, and the Camera app.

The idea for Photos is that with images containing people or objects, the AI could identify these and give the user the ability to order them to be cut out and pasted to another image (or document, or wherever).

With Paint, the idea might be to usher in generative AI, meaning you could ask the app to create something specific (in the same vein that the Bing chatbot can knock up a composition of an image when given a brief).

For the Snipping Tool, the idea is to introduce OCR (optical character recognition) which could allow the tool to pick up on text in screenshots, facilitating the extraction of those words to the clipboard. The Camera app may also get a similar OCR trick to pull out text from photos.

Bowden underlines that these ideas are still at the experimental stage, and it’s not clear when they might come to Windows 11.


Analysis: Windows 11 or Windows AI?

The fact that these capabilities are experimental suggests that they may not be coming to Windows 11 in the near future (if they ever do – this is just speculation, after all). Who knows, though – Microsoft could have a lot of the groundwork in place (for example, the image creation functionality already in Bing AI), and it could happen faster than we think.

Certainly, these are going to be very handy features if they do arrive in Windows 11. They’ll cut corners in impressively swift ways – pulling text out of photos more or less instantly, for example, and cutting out a person from an image in the blink of an eye (a painstaking task when performed manually, of course).

So, what’s the catch? Well, apart from the rumored nature of these AI features, Bowden does observe that they may have hardware requirements in some cases. That could include the presence of a VPU or vision processing unit, for example (a capability set to be built into Intel’s next-gen Meteor Lake CPUs, which speeds up AI-based tasks).

All of the above-mentioned features sound realistic inclusions for Windows 11 at some point, beefing up the operating system’s core apps considerably, and fitting neatly alongside Windows Copilot in making the OS a more AI-driven experience all-round.

If this rumor makes one thing clearer, it’s Microsoft’s ambition to infuse Windows 11 with AI in every corner – or maybe next-gen Windows, which as Bowden observes, will incorporate AI more ‘significantly.’ Perhaps that’ll be Windows 12, as rumored, or maybe the next incarnation of the OS could called Windows AI, even?

You might also like

TechRadar – All the latest technology news

Read More

Microsoft’s latest trick for Edge could be the closest Windows 10 users get to Copilot AI

Microsoft’s Edge browser has a new version which brings in some fresh features including the ability to detach the sidebar and move it onto the desktop for Windows 10 users.

This ability has been introduced with version 116 of Edge, as spotted by Ghacks, and it comes alongside the usual bug fixes and smoothing out of performance issues.

The Edge sidebar normally nestles on the right-hand side of the browser, but now, those on Windows 10 can pop it out of the browser window, and place it on their desktop.

The idea is to facilitate a “side-by-side experience” with the sidebar and any Windows 10 app, with the feature remaining present on the desktop, even if the Edge browser itself is closed.

So, this is kind of like having two taskbars on your desktop, if you will, with one of them being Edge-specific.

The Edge sidebar offers quick access to various bits of functionality, such as pinned websites, and Microsoft’s tools like Bing AI.


Analysis: Substitute Copilot – at least in a small way

This is a useful option that’s opt-in as Microsoft makes clear, so if you’re not interested in having the Edge sidebar on your Windows 10 desktop, you’ll never need to bother with it. For those who do want access to its features independently of the browser window, it’s clearly a handy choice to have.

Indeed, when you remember that Microsoft’s Copilot AI is only coming to Windows 11, this is actually a way of getting something a little like this on Windows 10. We’ve already seen that Microsoft plans to incorporate Copilot into the Edge sidebar, after all, so you’ll be able to deploy this on the desktop, in the same vein as Windows Copilot.

Granted, the functionality of Copilot for Edge will be nowhere near as useful as the full version of Copilot – which theoretically will be able to change all manner of Windows settings in the blink of an eye – but it’s something.

And Microsoft is going to work on adding “additional features and options” to the sidebar with future incarnations of Edge, as you might imagine. The sidebar isn’t going away, in short, for Windows 10 or 11 users, and is seemingly a key part of Microsoft’s ambition to make Edge one of the best web browsers out there.

You might also like

TechRadar – All the latest technology news

Read More

The Apple Vision Pro could pack in more storage than the iPhone 15

We know that the Apple Vision Pro isn't going to be available to buy until 2024, but we're learning a little bit more about the specs of the device through leaks from early testers – including how much on-board storage the augmented reality headset might pack.

According to iPhoneSoft (via 9to5Mac), the Vision Pro is going to offer users 1TB of integrated storage as a default option, with 2TB or 4TB a possibility for those who need it (and who have bigger budgets to spend).

Alternatively, it might be that 256GB is offered as the amount of storage on the starting price Vision Pro headset, and that 512GB and 1TB configurations are the ones made available for those who want to spend more.

This information is supposedly from someone who has been given an early look at the AR device, and noticed the storage space listed on one of the settings screens. It's more than the standard iPhone 15 model is expected to have – if it sticks with the iPhone 14 configurations, it will be available with up to 512GB of storage.

Plenty of unknowns

It does make sense for a device like this to offer lots of room for apps and files, and it might go some way to explaining the hefty starting price of $ 3,499 (about £2,750 / AU$ 5,485). Watch this space for more Vision Pro revelations as the launch date gets closer.

While the Apple Vision Pro is now official, there's still a lot we don't know about it – and it may be that we won't find out everything until we actually have the headset in our hands and are able to test it fully.

There have been rumors that two more Vision Pro headsets are in the pipeline, and that some features – such as making group calls using augmented reality avatars – will be held back until those later generations of the device go on sale.

We're also hearing that Apple might not be planning to make a huge number of these headsets, so availability could be a problem. Right now it does feel like a high-end, experimental device rather than something aimed at the mass market.

TechRadar – All the latest technology news

Read More

This bonkers Apple patent could solve one of VR’s biggest problems

Apple might have found a wild solution to VR’s prescription lens problem; liquid lenses.

VR headsets and glasses don’t usually mix well. Often they sit too close to your face for glasses to fit in front of your eyes, and the solutions deployed by headset designers are a mixed bag – some package in optional spacers that make room for specs like the Oculus Quest 2, while others include a prescription lens attachment (but you need to buy lenses for it at an added cost) like the Apple Vision Pro, and a few do nothing at all.

This has resulted in some glasses wearers feeling like VR isn’t accessible to them, but that might change if Apple’s latest patent comes to one of its headsets.

According to the patent granted in the US (via Apple Insider) Apple has created a design for an “electronic device with liquid lenses.”  The document describes a “head-mounted device” (sounds like a VR headset) with “tunable liquid lenses.” You can read the patent for the full details, but the TL;DR is that electronic signals sent to the lenses will deform the liquid in them and alter the refractive index of the lenses. 

This should allow the liquid lenses to correct a wide range of eyesight issues without the need for any accessories. What’s more, the correction is calibrated by the headset’s eye-tracking system.

Apple’s patent also states that it could apply to a “pair of glasses.” We can’t read too much into patent wordings, but this could hint at the Apple AR glasses that Apple apparently also has in development.

When will we get liquid lenses? 

Apple logo seen through a pair of glasses

Apple’s liquid lenses could bring VR and AR into focus (Image credit: Shutterstock / Girts Ragelis)

As with all patents we need to note that there’s no guarantee that we’ll ever see these liquid lenses appear in a real headset – one that’s made by Apple or otherwise. While the design exists in theory, it might be less than practical to add liquid lenses to a commercially available headset – either from a design or cost perspective. Alternatively, Apple might develop a different solution to VR’s prescription problem.

What’s more, even if liquid lenses do appear in an Apple headset you or I could pick up off the shelf there’s no telling when that will happen.

It’s probably an impossibility for the first-generation Vision Pro to launch in early 2024, and we’d be surprised if it appeared in the second-generation headset that rumors predict will appear sometime in the next few years. Instead, it seems far more likely we’d see liquid lenses in the third-generation model (assuming we see them at all) in half a decade or so – as this would give Apple plenty of time to hone the design.

You might also like

TechRadar – All the latest technology news

Read More

Google Chrome’s latest AI feature could be bad news for bloggers

Google continues to sprinkle AI into its product line with a new feature coming to its Search Generative Experience (SGE), where Google’s Chrome web browser will summarize the articles you’re reading. Currently, SGE already summarizes your search results so you don’t have to scroll forever, but with this new feature, you’ll get a little more help after you’ve clicked the link.

According to a Google blog post announcing the new feature, we won’t see the new feature right away, which Google has named “SGE while browsing”. The generated summary will begin rolling out next week Tuesday as an “early experiment” within the opt-in Search Labs, a program for people to experiment with early Google Search experiences and share feedback. Interestingly, it will be available on mobile before the Chrome browser on desktop, so keep your eyes peeled while you’re Googling on Android or iOS. 

Google generative AI in action

(Image credit: Google)

As you can see, the little pop-up appears as you’re scrolling through a blog page or article, and you’ll be able to see what Google’s tool thinks are the key points of the page. If you click a highlighted point, you’ll be taken down to that paragraph in the article.

The Verge notes that the feature “will only work on articles that are freely available to the public on the web,” so you won’t see it on websites or articles that are behind a paywall.

There are a few other smaller features that will be introduced as well via the SGE, like being able to hover over certain words and see definitions or diagrams (mostly for scientific, economic or historical topics).

Should Bloggers be worried? 

This feature could certainly be super helpful, especially if you’re looking for concise information very quickly. However, it may be bad news for the people writing the content and I don’t think Google has considered that.

If you’ve spent time and energy to really flesh out your article, give your topic context and personality and people are just given summaries and skip over that, it may be discouraging. Especially if you’re writing about sensitive or serious topics, if the generated summaries are leaving out crucial information people may only go with the presented points and leave behind something important or useful (like mixing cleaning products or medication information).

Content creators may not be happy about the new change and the way it might CEO Sundar Pichai said that “over time this will just be how Search works” so, I guess we’ll have to get used to it. 

TechRadar – All the latest technology news

Read More

This new YouTube Music feature could be the best way to discover new artists

YouTube Music is implementing a Samples tab on mobile in an effort to introduce new artists to potential fans via “short-form video segments”. Basically, it’s TikTok.

The announcement states Samples will have their home in the bottom navigation bar in between Home and Explore. Selecting it launches the personalized feed where the algorithm will display “the latest release from an up-and-coming artist or a deep cut from a legacy [musician]” the website thinks you would enjoy. Each track will be accompanied by a 30-second video clip. Swiping up on your phone screen, as you probably guess, skips to the next song. 

On the surface, Samples sounds similar to the Supermix and Discover playlists already present on YouTube Music. In a recent Engadget report, YouTube Music product manager Gregor Dodson claims the algorithm for Samples is different. Apparently, the new feature is a mix between Supermix and Discover, highlighting musicians you may know while also throwing in clips you might not have seen before.

Right now, you may be rolling your eyes at the fact that yet another popular social media app is copying TikTok’s endless feed. However, considering YouTube Shorts have proven to be very popular with its user base, plus the near-infinite amount of songs on the platform, adding the same feature to YouTube Music just makes a lot of sense.

Music demo

We managed to get our hands on Samples, and we have to admit, it’s pretty cool. It’s fun to see music videos you may not normally watch to then discover an awesome band you never heard of before. Be aware each snippet will loop endlessly. They won’t change automatically. To watch the next entry, you’ll have to manually swipe up on the screen.

On the side, you’ll have a series of buttons for liking songs, adding them to a playlist, sharing your favorites with friends, or using them in a YouTube Short. Tapping the three dots on the bottom right opens a menu leading to an extra set of tools. As you can see in the image below, users will be able to download songs (assuming you’re a YouTube Premium subscriber) or check out the musician’s profile.

YouTube Music Samples tools

(Image credit: Future)

Available now

If it wasn’t already clear, Samples is a free addition. You don’t need to subscribe to the Premium plan. Just make sure you have the latest version of YouTube Music on your mobile device. It’s currently rolling out to all users across the globe so keep an eye out for the patch when it arrives. 

There are plans to expand the tech to other parts of the platform. Details for future expansions are unknown at the time of this writing.

Melding music with an infinite feed seems like a growing trend. Spotify implemented similar tech when it redesigned its mobile app. And TikTok is going a different route by preparing its own music streaming service. To be honest, we're a little curious to see how long it’ll be until we see Tidal begin supporting a scrolling feed.

While we’re on the topic, check out TechRadar’s list of the best music streaming services for 2023.  

TechRadar – All the latest technology news

Read More

WhatsApp is about to get its first AI trick – and it could be just the start

WhatsApp is taking its first steps into the world of artificial intelligence as a recent Android beta introduced an AI-powered, sticker generation tool

Revealed in a new report from WABetaInfo, a Create button will show up in chats whenever some app testers open the sticker tab in the text box. Tapping Create launches a mini-generative AI engine with a description bar at the top asking you to enter a prompt. Upon inputting said prompt, the tool will create a set of stickers according to your specifications that users can then share in a conversation. As an example, WABetaInfo told WhatsApp to make a sticker featuring a laughing cat sitting on top of a skateboard, and sure enough, it did exactly as instructed. 

WhatsApp sticker generator

(Image credit: WABetaInfo)

It’s unknown which LLM (large language model) is fueling WhatsApp’s sticker generator. WABetaInfo claims it uses a “secure technology offered by Meta.”  Android Police, on the other hand, states “given its simplicity” it could be “using Dall-E or something similar.” 

Availability

You can try out the AI tool yourself by joining the Google Play Beta Program and then installing WhatApp beta version 2.23.17.14, although it’s also possible to get it through the 2.23.17.13 update. Be aware the sticker generator is only available to a very small group of people. There’s a chance you won’t get it. However, WABetaInfo claims the update will be “rolling out to more users over the coming weeks,” so keep an eye out for the patch when it arrives. No word on an iOS version. 

Obviously, this is still a work in progress. WABetaInfo says if the AI outputs something that is “inappropriate or harmful, you can report it to Meta.” The report goes on to state that “AI stickers are easily recognizable” explaining recipients “may understand when [a drawing] has been generated”. The wording here is rather confusing. We believe WABetaInfo is saying AI content may have noticeable glitches or anomalies. Unfortunately, since we didn’t get access to the new feature, we can’t say for sure if generated content has any flaws.

Start of an AI future

We do believe this is just the start of Meta implementing AI to its platforms. The company is already working on sticker generators for Instagram and Messenger, but they’re seemingly still under development. So what will the future bring? It’s hard to say. It would, however, be cool to see Meta finally add its Make-A-Scene tool to WhatsApp.

It’s essentially the company’s own take on an image generator, “but with a bigger emphasis on creating artistic pieces.” We could see this being added to WhatsApp as a fun game for friends or family to play. There’s also MusicGen for crafting musical compositions, although that may be better suited for Instagram.

Either way, this WhatsApp beta feels like Meta has pushed the first domino of what could be a string of new AI-powered features coming to its apps.

TechRadar – All the latest technology news

Read More

Amazon’s new AI review summarizer could help you empty your pockets even faster

Have you ever been shopping on Amazon, but found yourself too lazy to read the user reviews at the bottom of a product listing? Well, you’re in luck because Amazon has recently implemented a generative AI to its platform that will summarize reviews.

The company states the AI tool will offer short paragraphs “on the product detail page” highlighting key features as well as overall “customer sentiment”. Customers can quickly scan the short block of text to get an idea of whether a product is good or not instead of having to read dozens of reviews. Amazon states in its announcement you can direct the AI a bit by having it focus on specific “attributes.” Say you want a smart TV that’s easy to use. Users can select the “Ease of use” tab to have the summarizer specifically talk about that attribute or something else like its performance is while streaming content.  

Work in progress

Unfortunately, the AI feature was unavailable to us as we were excluded from the rollout, but The Verge had access. In their report, The Verge claims it saw the tool show up on listings for “TVs, headphones, tablets, [plus] fitness trackers.” It isn’t very consistent either. They state the summarizer is available on the Galaxy Tab A7, but not the Galaxy Tab A8. Also, it appears Amazon’s AI heavily favors writing positive content, as it spends “less time on the negatives.”

We reached out to Amazon with several questions about the new tool, including if there will there be a desktop version and if the company plans on providing links directing users to the AI's source reviews. Google’s SGE tool does this for the generated content it produces. It’d be nice to see sources in the paragraph. However, Amazon has nothing more to share at the moment.

Analysis: Remaining skeptical

Amazon has been dabbling in AI for a while now. Back in May, Amazon listed a job listing for a “machine learning focus engineer,” revealing the company is looking for someone to help develop an “interactive conversational experience” for its search engine. We could see the Amazon search bar one day offer a ChatGPT-like experience where you talk with the AI when looking for a product.

It would be wrong of us not to add a little asterisk to all this AI talk. As you may know, generative AIs are known to “hallucinate”, which is to say, they sometimes provide inaccurate information. It’s gotten to the point some experts believe this problem will never be fixed. So read the summarizer’s text with several grains of salt. As it turns out, you just can’t beat good old-fashioned human opinion – like the kind TechRadar provides every single day.

Labor Day is coming up and that means savings. Be sure to check out TechRadar’s guide for Amazon’s Labor Sale for 2023. Price cuts for certain electronics are already live.

TechRadar – All the latest technology news

Read More

Following Bing AI, Google could bring AI writing tools to Chromebooks

Google is supposedly preparing to introduce an AI-aided feature that will help users write, rewrite and edit text – and it could be coming to Chromebooks.

Google is putting in major efforts in this direction, already having announced Project IDX at its I/O conference earlier this year

Project IDX is a program that is currently in a preview stage that will help developers with all kinds of actions, from code development to previewing their projects on different platforms, and is enhanced with AI. Throughout I/O 2023, Google explained how it was adding artificial intelligence capabilities into its products and services in the near future.

Google's generative AI tools

There are already a range of AI-charged writing features incorporated into Google products. 

In Gmail and Google Docs, you may have seen “Write for me” or “Help me write” which give you ideas and suggestions to help you write for professional purposes. On mobile devices, Google has also added a “Magic Compose” option in Google Messages to revise a reply you’ve written, or to draft a reply based on the context of your ongoing conversation.

Two phone screens drawn in a cartoony style, the space around the phones and screens are covered in messages, drawings of file types and emojis

(Image credit: Google)

Rumblings around Google's new works

As for this latest rumor, 9to5Google suggests that there are five codenames for it at present, including “Orca,” “Mako,” and “Manta.” Apparently, “Orca” will appear in the ChromeOS right-click menu when you are editing a piece of text. After you select the text and click on Orca (whatever it looks like in the version it’s presented in), Orca will prompt the “Mako” UI to appear in a “bubble.” 

The Mako feature will then give you three choices for what it can do with your text, according to inspection of the code. The first is that you can “request rewrites” for the selected text and possibly give you some options of AI-revised versions. The second option will let you choose from a list of “preset text queries,” which 9to5Google proposes will suggest styles to rewrite your text. The final option will let Mako swap your text for a version that it suggests into whatever program, app, or page you’re working in. 

When you ask Orca to open a Mako suggestion bubble, then the Manta UI will send your original text input to Google’s servers, and then receive the generated suggestion to present to you. 

This means that the process of reworking your text doesn’t happen on your local ChromeOS machine. Presumably like the Magic Compose feature, you ‘ll have to provide clear consent to send your writing to the Google servers in this way.

9to5Google found that these mechanisms seem to be embedded into an upcoming version of ChromeOS, assuming it will show up in a future update. This will mean that it might be possible for the Orca UI to show up in nearly any app on your ChromeOS device (such as any of the best Chromebooks). It suggests this new writing assistant might be in the 118 ChromeOS update, due in mid-October. We don’t know this is the case definitely, and if you’re interested, be on the lookout for more intel from Google itself. 

Asus Chromebook

(Image credit: Future)

Possible Chromebook X exclusive?

There are also signs that Orca/Mako/Manta might only be incorporated into Chromebook X devices. Chromebook X is set to be a line of high-end laptops and tablets that was reported earlier this year. As Chromebook X will have higher spec requirements than existing Chromebooks, it could mean that when this feature is rolled out, it may not be available for all existing ChromeOS devices. 

This would be a pity and maybe a missed opportunity, in my opinion, and I hope that this won’t be the case. Microsoft has also recently debuted an AI assistant writing feature for its Bing AI chatbot in the Edge browser, and as far as we know, that won’t require any hardware beyond that which can run the latest versions of Windows 11 and Edge. 

Based on my experience of Bard, it still has a way to go to match ChatGPT (another AI tool, which Microsoft’s Bing AI is based on) in terms of writing (and rewriting) ability. We’ll see how widespread the availability of this AI-assisted tool is, but the more users that have access to it, the more it can improve.

TechRadar – All the latest technology news

Read More

Apple Vision Pro’s killer app could be… Windows XP?!

When the Apple Vision Pro, the powerful and ultra-expensive VR headset, was shown off by CEO Tim Cook at WWDC 2023, few (if any) of us expected one of the most interesting showcases for the new tech to be running the ability to run a Windows operating system by arch nemesis Microsoft – especially one that was first released back in 2001. But you know what? It actually is.

As 9to5Mac reports (via iPhoneSoft), developers working on an early version of the visionOS operating system that the Vision Pro will run on, have managed to get an emulator running with a working version of Windows XP.

In a video posted on X, the social media network formerly known as Twitter, which you can see below, Windows XP is shown loading in a big floating window in a lounge.

See more

Room with a view

While Windows XP is regarded as one of the better versions of Microsoft’s operating system, the idea of it looming over you as you sit on your couch may seem like some sort of dystopian nightmare, but this is actually pretty cool.

Sure, it’s unlikely that I’d want to fire up Windows XP to play some Minesweeper on a virtual 100-inch screen, this is an exciting demonstration on what could be possible for Vision Pro.

The developers are working on UTM, an emulator that brings non-Apple software to iPhones, Macs and now, it seems, the Vision Pro.

This emulation isn’t yet perfect – there isn’t a way to control Windows XP when it’s running – but the team has time to work on that before the Vision Pro’s official launch early next year.

And, while Windows XP is shown off in this video, it does suggest that this could mean other operating systems could come to Vision Pro. This would open up huge possibilities, as you’d be able to run full programs and games on the headset.

Apple Vision Pro

(Image credit: Apple)

Killer app? Perhaps

One of the major questions many people – including myself – had when Apple showed off the Vision Pro, was what is the headset actually for? We were shown some concept videos of people making video calls and watching movies using the headset, but nothing that really justified the huge $ 3,499 (around £2,815 / AU$ 5,290) price tag that it will launch with.

What we need is a killer app that makes the Vision Pro a must-by. So far, we’ve not had that, but an app that allows an almost unlimited amount of applications could be the key – and would also showcase Apple’s vision for ‘spatial computing’, which is how the company refers to the tech powering the Vision Pro – which includes hardware such as the same M2 chip found in the best MacBooks, and the new R1 chip.

The UTM app is certainly exciting, but I wouldn’t get too excited just yet. As you may imagine, Apple won’t be too keen on people running non-Apple software on the Vision Pro, so don’t expect installing UTM on the headset to be as straightforward as downloading it from the built-in app store.

It would be a shame if Apple hobbled the Vision Pro’s potential by forcing a walled-garden approach to apps, like on the iPhone, where you can only officially install apps from Apple’s own store, unlike the more open approach on Macs and Windows laptops

If Apple is serious about the Vision Pro being a productivity machine and the dawn of ‘spatial computing’, then it’s going to have to be willing to give up some control – and it may not want to do that.

TechRadar – All the latest technology news

Read More