Page 50 of 104

AI could be taken over by a few multinationals, warns UN

The United Nations’ High-Level Advisory Body on Artificial Intelligence, created last year to address AI governance issues, has made seven recommendations to address the risks with this technology in its first report, just published.

The document, entitled Governing AI for humanity, highlights the importance of creating a global dialogue — the European Union is one of the few to have acted with its EU AI Act — by building a global fund that addresses differences in capacity and collaboration and exchange standards.

And it warns of the dangers that AI can pose if it can be controlled in the market by only a few multinationals. “There is a risk that technology could be imposed on people without them having a say in how it is used,” it says.

Google wins landmark EU antitrust battle, easing legal pressures

Google landed a major antitrust victory on Wednesday as the EU’s Court of Justice annulled a $1.7 billion (€1.49 billion) antitrust fine against the search giant.

The EU court stated that the European Commission failed to fully consider all relevant factors regarding the duration of the abusive contract clauses. The court, however, upheld most of the Commission’s findings.

“By today’s judgment, the General Court, after having upheld the majority of the Commission’s findings, concludes that that institution committed errors in its assessment of the duration of the clauses at issue, as well as of the market covered by them in 2016,” the court said in a statement.

The European Commission, in its 2019 ruling, had accused Google of exploiting its market dominance by restricting websites from using ad brokers other than its AdSense service for search ads.

This would come as a relief for Google, which is facing other legal challenges. Just last week, the European Data Protection Commission (DPC) opened an inquiry into its use of personal data.

In August, a US District Court ruled that the tech giant is a monopoly, accusing it of leveraging its dominance in the online search market to hinder competition. A separate trial centered on its advertising business also commenced this month.

Meanwhile, in a separate ruling, Qualcomm failed to get the EU penalty imposed on it overturned, highlighting contrasting outcomes for the two major players in their ongoing regulatory battles with European authorities.

Implications for Big Tech

The Google ruling could influence future antitrust investigations and enforcement in the region, potentially leading to a more balanced and competitive environment, though the long-term impact remains uncertain.

“In this instance, the ruling is favorable for Big Tech as it signals a shift toward a more balanced approach to antitrust enforcement, taking into account both market impact and consumer welfare,” said Thomas George, president of Cybermedia Research. “This victory could foster a competitive environment where the constant threat of sanctions does not stifle growth and innovation.”

However, the court’s citation of errors in the Commission’s investigation, including issues with its definitions, suggests that a likely outcome is the Commission adopting a more cautious approach moving forward.

 “While this is a significant win for Google, it is worth noting that the court largely agreed with the arguments against the company, and the annulment was largely driven by the commission’s failure to build a strong case,” said Mayank Maria, vice president of Everest Group.

Maria added that this suggests there may not be a significant shift in the bloc’s approach toward Big Tech in the near future, even as new leaders take charge of two key roles — the antitrust chief and the digital chief — responsible for regulating Big Tech practices in the EU. Earlier this week, the European Commission appointed a new team to lead the institution for the next five years.

Qualcomm ruling in contrast


In Qualcomm’s case, the company secured a slight reduction of its EU antitrust fine, lowering the penalty to $266 million (€238.7 million) from the original $270 million (€242 million), but the court dismissed its other claims.

The fine, handed down by the European Commission in 2019, was based on claims that Qualcomm engaged in predatory pricing from 2009 to 2011, selling chipsets below cost to undercut British software firm Icera, now owned by Nvidia.

Compared to the Google ruling, Qualcomm’s failure to successfully appeal its fine highlights a different set of challenges, underscoring the varied regulatory issues impacting different areas of the technology sector, according to George. “As for its relevance to the semiconductor and chip markets, this ruling reinforces the substantial restrictions these markets will continue to face, particularly concerning competitive practices,” George said. “Competition within the semiconductor space could drive businesses to address issues like predatory pricing and stricter antitrust enforcement.” 

With genAI models, size matters (and smaller may be better)

As organizations continue to adopt generative AI (genAI) tools and platforms and explore how they can create efficiencies and boost worker productivity, they’re also grappling with the high costs and complexity of the technology.

The foundation of genAI and AI in general are language models, the algorithms and neural networks that power chatbots like OpenAI’s ChatGPT and Google’s Bard. The most popular and widely used models today are known as large language models (LLMs).

LLMs can be massive. The technology is tied to large, diverse troves of information and the models contain billionssometimes even trillions — of parameters (or variables) that can make them both inaccurate and non-specific for domain tasks or vertical industry use.

Enter small language models (SLMs), which have gained traction quickly and some even believe are already becoming mainstream enterprise technology. SLMs are designed to perform well for simpler tasks; they’re more accessible and easier to use for organizations with limited resources; they’re more natively secure because they exist in a fully self-manageable environment; they can be fine-tuned for particular domains and data security; and they’re cheaper to run than LLMs.

According to Ritu Jyoti, a group vice president of IDC’s AI research group, SLMs are well suited for organizations looking to build applications that can run locally on a device (as opposed to in the cloud) and “where a task doesn’t require extensive reasoning or when a quick response is needed,” Jyoti said.

Conversely, LLMs are better suited for applications that need orchestration of complex tasks involving advanced reasoning, data analysis and a better understanding of context.

SLMs can be built from scratch using open-source AI frameworks, which means an organization can create a highly customizable AI tool for any purpose without having to ask for permission, it can study how the system works and inspect its components, and it can modify the system for any purpose, including to change its output.

Open-source affords more freedom, customization

Dhiraj Nambiar, CEO of AI prototype developer Newtuple Technologies, said SLM adoption is growing because they can be fine-tuned or custom trained and have demonstrated “great performance for a narrow range of tasks, sometimes comparable to much larger LLMs.”

For example, he said, there are SLMs today that do “a great job” at optical character recognition (OCR) type tasks, and text-to-SQL tasks. “Some of the open-source ones are showing comparable performance to the LLMs,” Nambiar said.

In fact, the most popular SLMs today are open-source, IDC’s Jyoti said. They include:

The most popular non-open-source SLMs (which are proprietary and not freely available for public use) include:

“These models are typically used within specific organizations or offered as part of commercial services, providing advanced capabilities while maintaining control over their distribution and use,” Jyoti said.

An AI model infers from inputs the outputs it will generate, such as predictions, content, recommendations, or decisions that can influence physical or virtual environments. Different AI systems vary in their levels of autonomy and adaptiveness after deployment.

In the simplest of terms, a small language model (SLM) is a lightweight genAI model. The “small” in this context refers to the size of the model’s neural network, the number of parameters and the volume of data on which it is trained, according to Rosemary Thomas, a senior technical researcher in the AI lab at Version 1, a management consulting and software development firm. She said while some SLM implementations can require substantial compute and memory resources, several can run on a single GPU and have more than 5 billion parameters.

Those include Google Gemini Nano, Microsoft’s Orca-2–7b and Orca-2–13b, Meta’s Llama-2–13b, and others, Thomas noted in a recent article.

Adoption of SLMs is growing, driven by the need for more efficient models and the speed at which they can be trained and set up, according to Thomas. “SLMs have gained popularity due to practical considerations such as computational resources, training time, and specific application requirements,” she said. “Over the past couple of years, SLMs have become increasingly relevant, especially in scenarios where sustainability and efficiency are crucial.”

When compared with LLMs, the key difference lies in scale. Larger models are trained on vast amounts of data from diverse sources, making them capable of capturing a broad range of language patterns, where SLMs are more compact and trained on smaller, often proprietary. datasets. That allows for quicker training and inference times.

LLMs also require more computational resources and longer training times. “This makes SLMs a more practical choice for applications with limited resources or where quick implementation is needed,” Thomas said.

Though LLMs shine in tasks like content generation, language translation, and understanding complex queries small models can achieve comparable performance when correctly fine-tuned, according to Thomas.

“SLMs are particularly efficient for domain-specific tasks due to their smaller size and faster inference times,” she said.

Build or buy?

Organizations considering the use of an open-source framework to build their own AI models from scratch should understand that it’s both exorbitantly expensive and time consuming to fine-tune an existing model, according to Nambiar.

“There are a number of approaches in building your own AI model, from building it from scratch to fine-tuning an existing open-source model; the former requires an elaborate setup of GPUs, TPUs, access to lots of data, and a tremendous amount of expertise,” Nambiar said. “The software and hardware stack required for this is available, however, the main blocker is going to be the remaining components.

“..I highly recommend that for domain specific use cases, it’s best to ‘fine tune’ an existing SLM or LLM rather than building one from scratch,” he said. “There are many open-source SLMs available today, and many of them have very permissible licenses. This is the way to go about building your own model as of today. This broadly applies to all transformer models.” 

It shouldn’t be an all-or-nothing SLM strategy, said Andrew Brown, senior vice president and chief revenue officer at Red Hat. For one, training a single, general purpose AI model requires a lot of resources.

“Some of the largest models can require some 10,000 GPUs, and those models may already be out of date. In fact, research shows that by 2026, the cost of training AI will be equivalent to the US GDP, which is $22 trillion,” Brown said. “The average CIO doesn’t have a US GDP-level IT budget, nor do they have thousands of spare GPUs lying around. So, what’s the answer? Specialized, smaller AI models driven by open-source innovation.”

One of the big challenges in comparing costs across AI providers is the use of different terms for pricing — OpenAI uses tokens, Google uses characters, Cohere uses a mix of “generations,” “classifications,” and “summarization units,” according to Nambiar, whose company builds AI for business automation.

Nambiar settled on “price per 1,000 tokens” to evaluate varying prices.

Fine tuning an LLM for business purposes means organizations rely on AI providers to host the infrastructure. Nambiar said businesses should plan for a two-to-four month project based on both infrastructure and manpower. Costs typically start at $50,000 or more, Nambiar said.

Fine tuning SLMs will typically be more expensive, because if an organization hosts the opensource model, it will need to spin up the infrastructure – the GPU and/or TPU serves — as well as spend effort on fine-tuning and the labor costs. “Assume it will be more expensive than LLMs,” he said.

Clean data brings reliable results

Whether building your own or using a cloud-based SLM, data quality is critical when it comes to the accuracy. As with LLMs, small models can still fall victim to hallucinations; these occur when an AI model generates erroneous or misleading information, often due to flawed training data or algorithm. They can, however, more easily be fined tuned and have a better chance of being more grounded in an organization’s proprietary data.

As with LLMs, retrieval-augmented generation (RAG) techniques can reduce the possibility of hallucinations by customizing a model so responses become more accurate.

At the same time, due to their smaller size and datasets, SLMs are less likely to capture a broader range of language patterns compared to LLMs — and that can reduce their effectiveness. And though SLMs can be fine-tuned for specific tasks, LLMs tend to excel in more complex, less-well-defined queries because of the massive data troves from which they can pull.

“In short, SLMs offer a more efficient and cost-effective alternative for specific domains and tasks, especially when fine-tuned to use their full potential, while LLMs continue to be powerful models for a wide range of applications,” Thomas said.

Adam Kentosh, Digital.ai’s field CTO for North America, said it is extremely important with SLMs to clean up data and fine tune data stores for better performance, sustainability, and lower business risk and bias.  

AI initiatives have been sliding into the “trough of disillusionment,” something that could be avoided by addressing data quality issues, according to Kentosh.

By 2028, more than 50% of enterprises that have built LLMs from scratch will abandon their efforts due to costs, complexity and technical debt in their deployments.

“One of the biggest challenges we continue to face with existing customers is diverse data sources, even across common areas in software development,” Kentosh said. “For instance, most companies own two or more agile planning solutions. Additionally, there is almost zero consistency as it pertains to releasing software. This makes data preprocessing incredibly important, something that many companies have not been historically good at.”

Getting well curated, domain-specific data that works for fine tuning models is not a trivial task, according to Nambiar. “Transformer models require a specific kind of prompt response pair data that is difficult to procure,” he said.

And, once an organization decides to fine-tune its own SLM, it will have to invest in consistently keeping up with benchmarks that come from the state-of-the-art models, Nambiar said. “With every new LLM model release, the standards of inference capabilities go up, and, thus, if you’re creating your own fine-tuned SLM, you have to also raise the inference capabilities of this model, or else there’s no use case for your model anymore,” he said.

Brown said open-source AI models are not uncommon now, with industry giants such as Meta earlier this year championing the importance of its Llama model being open source. “That’s great news for organizations as these open-source models offer a lot of benefits, such as preventing vendor lock-in, allowing for a broad ecosystem of partners, affordability for the performance and more,” he said. “But unfortunately, none of that really matters if you don’t have the data scientists needed to work with the model.”

Brown described data scientists as unicorns right now — rare and often demanding the pay of a mythical creature, as well. “And rightly so,” he said.

Most organizations can only employ a handful of data scientists at best, whether due to a scarcity of qualified talent or the cost of employing them, “which creates bottlenecks when it comes to effectively training and tuning the model,” he said.

A move to hybrid?

CIOs, Brown noted, have long been moving away from monolithic technologies — starting with the shift from UNIX to Linux in the early 2000s. He believes AI is at a similar turning point and argues that a hybrid strategy, similar to that of hybrid cloud, is most advantageous for deploying AI models. While the large, somewhat amorphous LLMs are in the spotlight today, the future IT environment is 50% applications and 50% SLMs.

“Data lives everywhere, whether it’s on-premises, in the cloud or at the edge. Therefore, data by nature is hybrid, and because AI needs to run where your data lives, it must also be hybrid,” Brown said. “In fact, we often tell customers and partners: AI is the ultimate hybrid workload.

“Essentially, a CIO will have as many AI models as applications. This means that training needs to be faster, tuning needs to speed up and costs need to be kept down. The key to this challenge lies in open source,” he continued. “Just as it democratized computing, open source will do so for AI; it already is.”

Encryption is coming to RCS, protecting Android and iPhone

Now that Apple supports Rich Communication Services (RCS) messages on iPhones running iOS 18, the GSM Association (GSMA) has promised end-to-end encryption (E2EE) is coming to the standard, a move that should better protect communications between iPhone and Android devices.

The GSMA, which maintains the standard, is working to implement E2EE but hasn’t committed to a time scale. It announced the plans as it marked the launch of RCS support on the iPhone.

E2EE will be ‘next major milestone’ for RCS

“While this is a major milestone, it is just the beginning,” the GSMA said in a statement. “The next major milestone is for the RCS Universal Profile to add important user protections such as interoperable end-to-end encryption. This will be the first deployment of standardized, interoperable messaging encryption between different computing platforms, addressing significant technical challenges such as key federation and cryptographically-enforced group membership. Additionally, users will benefit from stronger protections from scam, fraud, and other security threats.”

RCS Universal Profile is an industry-agreed set of features and technologies the GSMA has standardized so RCS can be widely deployed in products and services. The most recent version, RCS Universal Profile 2.7, introduced support for more advanced messaging features, such as reactions, editing of sent messages, and improved message indicators.

Apple now supports RCS on iPhone

Apple has now adopted RCS within iOS 18, replacing the long-in-the-tooth SMS system for texts to Android devices. Messages between the platforms are much improved as a result — many users were annoyed that they couldn’t share high-resolution images, for example. 

However, the lack of E2EE is a glaring hole in messaging security; it means enterprise users will likely employ other messaging services for critical information. Apple’s own message system does support E2EE, but not when sharing with an Android device — hence, the colored bubbles to show you when a message is secure. You will know when you’re in an RCS chat with an Android user because you’ll see a small grey label that says RCS Message in the text field.

Other significant benefits

The GSMA promise of encryption in RCS is a welcome one. It will prevent carriers, messaging services, or other third parties with access to these communications from viewing the content of the texts you share or sharing that information for any reason.

Encryption on messages between platforms also promises other benefits, as noted last year by Caitlin Seeley George at Fight for the Future: “This move makes it possible for cross-platform messages to be end-to-end encrypted — a security feature that would protect a whole host of vulnerable groups, including pregnant people, LBGTQ+ people, activists, immigrants, and journalists.”

It is possible that Apple’s decision to introduce support for RCS might have helped it avoid its messaging service being declared a ‘Gatekeeper’ service under the EU’s DMA

Please follow me on Mastodon, or join me in the AppleHolic’s bar & grill and Apple Discussions groups on MeWe.

Apple delivers enterprise IT improvements for iPhone, iPad, and the Mac

While we’ve already focused on how new features in Apple’s latest operating system upgrades can help you get work done on an iPhone and on a Mac, under-the-hood enhancements for enterprise IT also made the cut. Here’s a rundown on what was introduced.

Apple’s changes for enterprise and education deployments lean toward declarative device management (DDM) as the company continues to navigate away from MDM profiles. They also include security tweaks to manage newly-introduced features such as iPhone Mirroring on Mac. 

The latter is interesting in its own right, as it means IT can prevent managed iPhones from being mirrored on any Mac, and any Mac can also be prevented from mirroring any iPhone. That makes sense, as it puts a barrier in place against abuse of that feature.

On the good ship DDM

Apple has flagged its move toward Declarative Device Management (DDM) since it first introduced support for this way of managing devices in 2021. This was originally only available to user-enrolled iOS devices, then Macs joined in. It is now supported across all the company’s products, including the Apple Watch and Apple TV. In 2022, Apple told us explicitly at WWDC that DDM will eventually supersede its own older MDM framework.

The beauty of DDM is that it allows IT to specify a desired state for devices. That means devices that do not occupy that “state” won’t get access, and the device doesn’t require continuous prompts from an MDM server.

Admins recognize that this makes for faster onboarding, better update management, improved device monitoring, better security and reduced network bandwidth usage. For most users, it just makes for a far nicer experience, bringing the convenience of consumer grade simplicity to managed enterprise devices. Everyone in the Apple MDM ecosystem I speak with has told me that DDM is the future for managed devices.

Michael Covington, Jamf vice president for portfolio strategy, last year noted the importance of this move. Pointing to DDM improvements at the time, he said: “Of course, the big announcement for those IT professionals responsible for managing devices is Apple’s improvements to Declarative Device Management. The new Software Update workflows demonstrate Apple’s commitment to iterating on the enhanced protocol….”

In this year’s biggest move toward that future, Apple’s new operating system software updates can now be managed entirely using DDM, replacing MDM profiles for software update restrictions, settings, and software update commands and queries.

A short summary of improvements for Apple admins

Of course, DDM is only one facet of device management, and while there are some unique differences between Mac and i-device platforms, Apple peppered the releases with new device management features:

  • MDM can manage which Safari extensions are allowed, always on or always off, and what websites they can access.
  • On supervised devices, organizations can disable a user’s ability to hide and lock apps.
  • IT can prevent VPN settings from being modified by apps.
  • A new MDM restriction can prevent the removal of an eSIM.
  • New features in Calculator, such as Math Notes, Math Notes keyboard, scientific mode, and unit conversions can be managed in MDM. (This tool is aimed at education IT.)
  • On a Mac, a new disk management configuration can be used to choose whether external or network storage is allowed or disallowed, or limit mounting to read-only volumes.
  • Also on a Mac, MDM can configure hardware MAC address instead of a private MAC address on managed Wi-Fi networks. MDM can also prevent system extensions from being disabled in Settings.
  • iPads gain an iPhone tool; IT can now use MDM to manage alternative marketplaces in regions in which those are supported.
  • visionOS 2 now supports Automated Device Enrolment in MDM

For the most part, Apple’s enterprise improvements seem designed to give IT additional power to harden security across managed devices while also working to prevent data leaks. Companies in which security policy is attenuated also benefit from a small but noteworthy improvement in which users with complex passwords no longer need to lock and unlock the device to see the keyboard.

Apple’s complete lists

There are other features and tools listed across Apple’s documents detailing the enterprise-focused content in the latest OS upgrades. Apple’s pages detailing these improvements are here:

What’s new for enterprise in macOS Sequoia

What’s new for enterprise in iPadOS 18

What’s new for enterprise in iOS 18

What’s new for enterprise in visionOS 2

While many of the changes described above may be of less interest to most users, they will be of huge significance to the ever-growing cadre of Apple admins who are seizing seats across the enterprise as the company’s market share across that sector continues to grow. Apple is in business, from the world’s biggest firms to SMBs, and its continued focus on empowering MDM teams to support that proliferation now seems to run deep. Apple knows that, unlike the main enterprise tech incumbent, its products aren’t renowned for causing business disasters. It knows it has a story to tell, and for many in business in a world after the Crowdstrike mess, it also knows it offers a viable and robust alternative

Please follow me on Mastodon, or join me in the AppleHolic’s bar & grill and Apple Discussions groups on MeWe.

The ultimate Windows app launcher

Ah, the morning ritual: Once you’ve booted up your computer, it’s time to open all those application windows you need to get work done. You might want to position them just so, especially if you use a variety of apps or rely on multiple displays in your workflow.

But wait: Computers are designed to automate repetitive tasks! Isn’t there be a better way to do this?

The answer is yes — and Microsoft’s new Workspaces PowerToy is how to make it happen. With a few fast clicks, the tool empowers you to pull up any specific group of application windows you need and put them exactly where you want them on your desktop.

It even supports multiple monitors for people who have especially complex desktop setups. If Microsoft were rolling this out as a feature in the next version of Windows, PC users everywhere would be excited! But it’s a free download today and most PC owners don’t even realize it’s available for the taking.

The Workspaces tool is part of Microsoft’s free PowerToys package for Windows 10 and Windows 11. You can download it from the Microsoft Store or Microsoft’s website. (Microsoft added the Workspaces utility in PowerToys version 0.84, released on Sept. 3. If you have an older version, you’ll want to launch “PowerToys” from the Start menu and use the “Check for updates” button on the General tab.)

Once you have PowerToys in place, the real fun begins.

Looking for more useful Windows tips and tricks? My free Windows Intelligence newsletter delivers all the best Windows tips straight to your inbox — along with free in-depth Windows Field Guides as a special welcome bonus!

Create your custom Windows workspace

Once you’ve opened up PowerToys at least once to confirm it has all the required permissions — you can start using the new Workspaces feature by pressing Windows+Ctrl+`. (That last character is known as the “grave accent” — it’s the key above the Tab key on your keyboard.)

(You can tweak this keyboard shortcut if you like. Launch PowerToys, select “Workspaces” in the sidebar, and change the activation shortcut.)

To save a workspace of windows, click the “Create Workspace” button. Now, create the “workspace” you want to use by opening windows, closing windows, and positioning everything where you want on your screen.

Workspaces window
Click the “Create Workspace” button to get started.

Chris Hoffman, IDG

PowerToys will remember everything — which applications are open, which windows are positioned where, and even which applications are minimized. When you’re done, click “Capture.” You’ll see a preview of what the Workspaces tool remembers.

You can easily edit this list: To remove an application you don’t want to associate with a particular workspace, just click the “Remove” button next to it.

When you’re happy with your workspace, give it a name using the “Workspace name” box at the top of your screen. Then, consider checking the “Create desktop shortcut” box to make your workspace easier to launch. Finally, click “Save Workspace.”

Workspace editing
Go through the list of remembered applications and click “Remove” to remove anything you don’t want launched.

Chris Hoffman, IDG

Launch a workspace of Windows apps

After you’ve saved your preferred setup, you have two options to launch it. The easiest is with that desktop shortcut you created just a moment ago. You just double-click the shortcut on your desktop and Workspaces will get to work, opening applications and positioning all those windows for you.

You can also open the Workspaces window by pressing Windows+Ctrl+` and then using the “Launch” button to launch any of your saved workspaces.

Workspaces launchers
Launch a workspace from the desktop shortcut or Workspaces window.

Chris Hoffman, IDG

Personally, I prefer the desktop shortcut. When you boot your computer and it’s time to work, you can quickly double-click the shortcut and have Workspaces take it from there.

(Don’t see the desktop shortcut? Right-click an empty spot on your desktop and select View > Show desktop icons to toggle desktop icons on and off.)

Workspaces launching
You’ll see a dialog box while PowerToys is launching and repositioning the windows.

Chris Hoffman, IDG

Tweak your workspace with advanced arguments

If you want to take your Windows workspace setup to even more advanced heights, you can dig into the feature’s app-specific command-line arguments. (These are why Microsoft is positioning PowerToys Workspaces as a tool for developers and not just regular business users!) When you’re creating or editing a workspace in the Workspaces window, you can click an application in the list, and you’ll see a “CLI arguments” field.

You can specify any command-line arguments here and PowerToys Workspaces will use them when launching the applications in that workspace.

Take Google Chrome, for example: You could specify a web address like https://theintelligence.com to launch that specific web page when your workspace opens. Or with a program like Microsoft Word or Excel, you could specify the path to a document or spreadsheet on your computer to launch it.

Workspaces CLI arguments
Click an application’s name and add arguments to the “CLI arguments” field, if you like.

Chris Hoffman, IDG

The available options depend on the command-line switches the application in question supports, so you might have to do some digging to uncover all the available options for any given program.

A Windows workspace warning

When you use PowerToys Workspaces, you could encounter one big problem: The system works best when you have no application windows open and you’re starting from a clean, empty workspace. If you already have some windows open, its behavior can be a little unpredictable.

Here’s the issue: Let’s say you have an app open and you try to launch it again from your Start menu. What happens? It depends on the application. For some applications — like Google Chrome and Microsoft Word — Windows will open an additional window. For other applications — like Spotify or Microsoft To Do — Windows will “reuse” the open application window, bringing it into focus.

That’s because PowerToys Workspaces functions by launching an application — just like you’d launch it from your Start menu — and then moving its window to a configured position on your desktop.

So, if you already have a Chrome browser window open and you launch a workspace that has a Chrome window, you’ll get a new Chrome window in the configured location. For other applications, PowerToys will reposition the current window.

For predictable behavior, consider closing all your open applications or just using PowerToys Workspaces when you first sign into your computer and have a clean desktop.

Other useful Windows workspace systems

Windows has lots of other ways to organize the apps you’re using, including the powerful Snap feature for organizing windows and Task View, which lets you organize groups of open windows into different virtual desktops. (Unfortunately, Workspaces isn’t integrated with those Task View virtual desktops.)

For more useful PowerToys, check out FancyZones, which provides extra organization powers for your open windows. And I recommend checking out the Always On Top PowerToy; it offers an easy way to make any window “always on top” of other windows, which is can be helpful in various workflows.

Want more? Dig into this list of 10 excellent PowerToys you should use on Windows. There are many powerful utilities included in this power-user-focused free download from Microsoft; this new Workspaces addition is merely the latest in a long list of worthwhile options.

Discover useful Windows tools like this one with my free Windows Intelligence newsletter. You’ll get all the best Windows tips straight into your inbox — and you’ll get free in-depth Windows Field Guides just for subscribing!

The anti-iPad-ification of Google’s Android tablets

If you’ve been following Android for long, you’re probably familiar with two seemingly at-odds realities about the role Google’s platform plays in the greater tech universe:

  1. Android has been the dominant mobile operating system (and all-around computing operating system) for ages now, thanks to its prominent and diverse presence on the phone side of things.
  2. At the same time, Android has long struggled to gain any meaningful foothold on the tablet end of the equation — with an endless-seeming array of about-faces and puzzling pivots in that area.

Plain and simple, while Android phones have always had their own distinctive identity and appeal, Android tablets have flipped and flopped more times than an inebriated penguin on ice skates. And between their lack of any consistent identity and the on-off nature of their development in general, that’s allowed Apple’s iPad to become the firmly entrenched de facto standard for slate-shaped devices while Android tablets have remained a relative afterthought, in comparison.

Now, though, it looks like Google may be ready to change that narrative — and change the very way we think about Android tablets and the role they play in our lives.

And there’s more than a passing connection to Apple’s own wildly successful tablet philosophy, as we’ll explore more thoroughly in a moment.

[Get fresh, original insight in your inbox with my free Android Intelligence newsletter. Three new things to know and try each Friday!]

Google’s Android tablet (re)reinvention

This latest pivot — yes, technically another one! — comes by way of a quietly published post on Google’s Android Developers Blog just ahead of last weekend.

The post announces a new developer preview of an unassuming-seeming feature called “desktop windowing on Android tablets.”

As El Googenthaal puts it:

Desktop windowing allows users to run multiple apps simultaneously and resize app windows, offering a more flexible and desktop-like experience. This, along with a refreshed System UI and new APIs, allows users to be even more productive and creates a more seamless, desktop-like experience on tablets.

Intriguing, no? I thought so. And then I started actually playing around with the under-development system on my own personal Pixel Tablet, and I quickly realized just how big of a deal this could actually be.

Translated into plain English, the system lets you easily take any app you’re seeing on an Android tablet and shrink it down into a resizable window — just like you’d do on a desktop computer.

Google Android tablet windowed apps
A desktop-like canvas on Android, with Google’s new windowed apps system for tablets.

JR Raphael, IDG

That means in addition to the phone-like ability to split the screen in half and view two apps at the same time, you can open up numerous apps and arrange ’em in any way you like — side by side, tiled, overlapping, you name it — with any number of apps visible at any given moment and windows taking up any size your precious heart desires.

It gives you a completely flexible canvas, in other words, and transforms the Android tablet from a mostly consumption-oriented gadget into a computer-like productivity machine. (Samsung, notably, has offered a similar option on its Android devices for some time now. But having such a system present at the actual Android platform level introduces a whole other level of seamless integration, universal compatibility, and all-around prominence — along with a more consistent experience when it comes to actual app interactions, given the operating-system-level connections.)

And even in its early, still-rough-around-the-edges shape, I can’t help but be blown away by just how significant of a change this seemingly small twist could shape up to be.

The ins and outs of Google’s Android tablet app windows

For context, with the new Android tablet desktop windowing setup — which is currently accessible (a) only on Google’s Pixel Tablet, (b) only if you opt the device in to the latest Android 15 beta update, and (c) only if you enable its developer settings and then find and flip the switch to a specific related setting — every app gains a new thin bar at the top-center of the screen.

When you want to move from the standard Android interface into the new windowed mode, you simply swipe down on that bar. And you can then place the app wherever you want and continue to move and resize it from there.

srcset="https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-gesture.webp?quality=50&strip=all 600w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-gesture.webp?resize=300%2C188&quality=50&strip=all 300w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-gesture.webp?resize=269%2C168&quality=50&strip=all 269w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-gesture.webp?resize=134%2C84&quality=50&strip=all 134w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-gesture.webp?resize=576%2C360&quality=50&strip=all 576w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-gesture.webp?resize=400%2C250&quality=50&strip=all 400w" width="600" height="375" sizes="(max-width: 600px) 100vw, 600px">
Shifting Android apps into flexible windows is as easy as swiping down once, then moving and resizing as needed.

JR Raphael, IDG

A few other interesting touches come into play once you’re in that Android tablet windowed mode:

  • A taskbar, similar to the one that can already be summoned on any large-sized Android device, appears in a permanent position at the bottom of the screen — providing an extremely desktop-reminiscent setup, with easy anytime access to your currently opened apps along with any pinned apps and your entire Android app drawer.
  • Each individual window gains its own title bar area, which contains an “x” for closing the window as well as an icon for maximizing the window and a dropdown for snapping the window into a series of standard sizes and positions.
  • And anytime you want to exit the windowed mode and go back to the standard Android interface, you can drag any window to the top of the screen to make that transition.
srcset="https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-exit.webp?quality=50&strip=all 600w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-exit.webp?resize=300%2C188&quality=50&strip=all 300w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-exit.webp?resize=269%2C168&quality=50&strip=all 269w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-exit.webp?resize=134%2C84&quality=50&strip=all 134w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-exit.webp?resize=576%2C360&quality=50&strip=all 576w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-exit.webp?resize=400%2C250&quality=50&strip=all 400w" width="600" height="375" sizes="(max-width: 600px) 100vw, 600px">
Pushing any Android app back to the top of a tablet’s screen is all it takes to exit the new windowed-apps mode.

JR Raphael, IDG

As for the broader implications, three points in particular jump out at me based on my time with the setup so far — the second of which ties into that tantalizing Apple connection at the start of this story:

1. The Android windowed tablet interface feels a lot like ChromeOS

As anyone who’s spent much time with Google’s other primary platform is likely to notice, this setup bears more than a passing resemblance to the main ChromeOS interface present on the company’s many Chromebook computers.

That’s true from the taskbar to the overall interface styling and also the ability to move an Android app between its phone-like state and a more large-screen-optimized interface simply by changing the width of its window.

srcset="https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-resize.webp?quality=50&strip=all 600w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-resize.webp?resize=300%2C188&quality=50&strip=all 300w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-resize.webp?resize=269%2C168&quality=50&strip=all 269w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-resize.webp?resize=134%2C84&quality=50&strip=all 134w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-resize.webp?resize=576%2C360&quality=50&strip=all 576w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-android-tablet-windowed-app-resize.webp?resize=400%2C250&quality=50&strip=all 400w" width="600" height="375" sizes="(max-width: 600px) 100vw, 600px">
Just like on ChromeOS, you can move between an Android app’s phone-style interface and its more desktop-like presentation simply by changing the app’s size.

JR Raphael, IDG

When Google first restarted its Android tablet strategy two years ago — just a few years after officially announcing it was done with focusing on tablet development entirely, in 2019 — I raised the question of how Android tablets and Chromebooks could coexist and make sense together without awkwardly overlapping and stepping on each others’ toes.

Early on, after all, the line between the two device types was easy to understand: Android was for smartphones, while ChromeOS was for laptops, desktops, and the most optimal Android-app-supporting tablet experience.

Once Google made it clear it intended to resurrect the native Android tablet as a prominent option, the narrative got much more muddled. But Google had an answer: When I asked for clarification in 2022, the company told me the difference ultimately came down to how you intended to use the product.

Go, go, gadget self-quoting machine:

In short, Android tablets are intended for “productive mobility” — with content consumption being the top priority and a bit of more complex productivity being an occasional add-on.

Chromebook tablets, on the other hand, are the exact opposite: They’re intended for “mobile productivity,” with the active work being the primary purpose and the more passive consumption being a pleasant side perk.

Now, it seems like Google’s working to rethink that narrative and make Android tablets about active productivity, too. So where does that leave Chromebooks — and how will any regular person figure out the differences between the two approaches and which type of device is most suitable for any given purpose?

That’s a question I’m not entirely convinced anyone is ready or able to answer.

2. This Android tablet expansion seems like a very deliberate response to the iPad’s shortcomings

While Apple may well wear the crown of the all-around tablet champion, one thing I’ve consistently heard from even the most enthusiastic Apple fanatics is that iPads are being held back by Apple’s intentionally limiting software.

In short, folks on the Apple side of the Great Tech Divide are hungry for a more powerful framework that brings desktop-like computing capabilities onto their otherwise capable devices.

Federico Viticci of MacStories assembled an impressively thorough list of such desires earlier this year, ranging from a fix for the iPad’s surprisingly rudimentary file management capabilities to an improvement for its “fractured mess” of multitasking shortfalls:

iPadOS’s multitasking … could be so much more. I think several iPad users (and I was guilty of this, too) have convinced themselves due to Apple’s pace of updates that we’ve reached the peak of what tablet multitasking should do with Split View and Stage Manager. But look outside Apple’s stance on iPadOS, and you see that is not the case. Once again, I’m not arguing for macOS features on the iPad; I’m saying that, if Apple wanted to, it could design innovative, high-performance, delightful tablet-first multitasking systems. Sadly, iPad multitasking tells a very different story.

Six Colors founder (and frequent Computerworld sister site Macworld contributor) Jason Snell has shared similar sentiments around the iPad’s productivity boundaries:

Professionals multitask. Professional tools should, too. This is an area where the iPad Pro fails its users. …

I have such an affinity for my iPad that I have wanted to integrate it into as much of my life as possible. And for an increasing number of tasks, I can. But for many others, I am eternally bumping up against the severe limitations of the platform.

At the risk of giving Google too much credit, it certainly feels like this Android tablet expansion is a direct reaction to the void Apple’s creating with those decisions. It positions the Android tablet as the anti-iPad, in a sense — the productivity-forward powerhouse of a gadget Apple refuses to offer. You want a tablet that can actually act like a computer and give you a desktop-caliber environment for going beyond passive consumption and genuinely getting stuff done? The iPad’s not gonna give it to ya, pal. But hey, we can.

Now, realistically, are throngs of Apple diehards suddenly gonna drop everything and abandon the ecosystem they’re so heavily invested in to pick up an Android tablet? That’s highly unlikely. But if Google can make some noise and win over even some segment of iPad owners — along with professionals who don’t presently rely on tablets for productivity — while simultaneously convincing Android phone owners to stay in its ecosystem and grab an Android tablet for their needs, well, that’d be a pretty significant victory to be able to claim.

And, perhaps just as important, it’d allow Android tablets to carve out a niche of their own beyond just being “iPads, but with Android” — something Google started to do on another front, with its attempt to frame the Pixel Tablet as a souped-up Smart Display last year, but never fully realized due to the ineffective implementation of that admirable-seeing idea.

(And anyway, the productivity angle is a far more powerful point of differentiation to be able to claim — given the door it cracks open into the lucrative and potential-packed world of business computing.)

3. Beyond just tablets, this type of concept could become very interesting for foldable phones, too

The timing of this launch is especially intriguing to me right now because I’m in the midst of spending some quality time getting to know Google’s latest foldable Android device — the awkwardly named but otherwise delightful Pixel 9 Pro Fold.

And as soon as I saw the news about the Android tablet windowing development, I couldn’t help but think: “Oooh. This could have huge implications here.”

Consider:

  • Google’s exerting a ton of effort to frame the Pixel 9 Pro Fold as being a best-of-both-worlds, business-ready phone and full-sized 8″ tablet in your pocket.
  • The main benefit of the Fold is the fact that you can use it like a regular phone most of the time but then pop it open at a second’s notice to unfold some serious productivity power — with the ability to view and even interact with two full-sized apps on-screen at the same time.
  • As I put it previously, this “strips away many of the standard limitations of using a phone-sized device and gives you a more desktop-like computing experience — which, suffice it to say, can be a massive asset when it comes to work-oriented, productivity-centric tasks.”

If a folding Android phone is essentially meant to bring a tablet-like experience into the palm of your hand on demand, then wouldn’t it make sense for this new Android tablet advantage to also be available within that environment?

It builds upon the same basic productivity-minded perks already present on a phone like this latest folding Pixel — with the more desktop-like experience and all the tools for being able to effectively work across different apps and processes without being limited to a simple 1:1 screen-split scenario.

Heck, it even involves the same exact exceptional taskbar already built into that fully unfolded part of the Pixel 9 Pro Fold’s interface:

srcset="https://b2b-contenthub.com/wp-content/uploads/2024/09/google-pixel-9-pro-fold-taskbar.webp?quality=50&strip=all 600w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-pixel-9-pro-fold-taskbar.webp?resize=289%2C300&quality=50&strip=all 289w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-pixel-9-pro-fold-taskbar.webp?resize=162%2C168&quality=50&strip=all 162w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-pixel-9-pro-fold-taskbar.webp?resize=81%2C84&quality=50&strip=all 81w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-pixel-9-pro-fold-taskbar.webp?resize=463%2C480&quality=50&strip=all 463w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-pixel-9-pro-fold-taskbar.webp?resize=347%2C360&quality=50&strip=all 347w, https://b2b-contenthub.com/wp-content/uploads/2024/09/google-pixel-9-pro-fold-taskbar.webp?resize=241%2C250&quality=50&strip=all 241w" width="600" height="622" sizes="(max-width: 600px) 100vw, 600px">
Google’s new Android taskbar, as seen in the inner “tablet” mode of a Pixel 9 Pro Fold phone.

JR Raphael, IDG

All in all, it feels like a no-brainer and like something Google would be crazy not to introduce into that type of device once the testing period is over and the feature actually launches. And that introduction would take a promising type of new mobile technology and make it even more powerful for that same business-minded, productivity-seeking sort of buyer.

Make no mistake about it: From tablets themselves to the foldables around ’em, there’s much more to this move than what we see on the surface. And even in its earliest form, the feature feels like a complete reinvention of the associated devices and what they’re able to accomplish.

If Google manages to implement and then market this effectively — a big “if,” admittedly, given the company’s past slip-ups and notoriously short attention span — this could be huge.

Watch this space.

Want even more Googley knowledge? Check out my free weekly Android Intelligence newsletter to get next-level tips and insight delivered directly to your inbox.

Sam Altman exits OpenAI commission for AI safety to create ‘independent’ oversight

OpenAI’s CEO Sam Altman has stepped away from his role as co-director of an internal commission the company created in May to oversee key safety and security decisions related to OpenAI’s artificial intelligence (AI) model development and deployment.

OpenAI’s Safety and Security Committee will become “an independent board oversight committee focused on safety and security” led by its new chair, Zico Kolter, director of the machine learning department of Carnegie Mellon University’s School of Computer Science, the company revealed in a blog post Monday. Kolter replaces the committee’s former chair, Bret Taylor, who also has departed.

Other members of the committee, which is chiefly aimed at overseeing the safety and security processes guiding OpenAI’s model development and deployment, remain: Adam D’Angelo, Quora co-founder and CEO; retired US Army General Paul Nakasone; and Nicole Seligman, former EVP and general counsel at Sony Corporation.

It was this committee, under Kolter’s leadership, that reviewed the safety and security criteria that OpenAI used to assess the “fitness” of OpenAI o1 for launch, as well as the results of safety evaluations for the model, according to the post. OpenAI o1 is the company’s latest family of large language models (LLMs) and introduces advanced reasoning that the company said exceeds that of human PhDs on a benchmark of physics, chemistry, and biology problems, and even ranks highly in math and coding.

More transparency, collaboration and monitoring on tap

OpenAI shared recommendations for the committee’s mission going forward: to establish independent governance for AI safety and security; enhance security measures; foster transparency about OpenAI’s work; collaborate with external organizations; and unify the company’s safety frameworks for model development and monitoring.

“We’re committed to continuously improving our approach to releasing highly capable and safe models, and value the crucial role the Safety and Security Committee will play in shaping OpenAI’s future,” said the post.

Indeed, AI safety, and OpenAI’s management of it in particular, is something that has become of great concern to various industry stakeholders and lawmakers.

Altman became a controversial figure soon after forming OpenAI, and his abrupt ousting from and subsequent return to the company late last year, and the behind-the-scenes deal making and shake-ups that occurred in the aftermath quickly led to infamy for the CEO, who has become a public face of AI.

Highlights in that journey included OpenAI securing a $13 billion investment from Microsoft, which uses OpenAI technology for its generative AI tool, Copilot, and breaking ideologically from Tesla’s Elon Musk, a controversial figure in his own right, who was one of OpenAI’s founding board members and investors. Musk ultimately sued OpenAI and Altman for breaching its founding mission.

The safety of OpenAI’s technology also has been called into question under Altman, after reports surfaced that the company allegedly used illegal non-disclosure agreements and required employees to reveal whether they had been in contact with authorities, as a way for it to cover up any security issues related to AI development.

Effect of the move as yet unknown

It remains to be seen what, if any, impact Altman’s stepping back from OpenAI’s safety board will have on AI governance, which is still in its infancy, noted Abhishek Sengupta, practice director at Everest Group.

However, it appears to be a sign that the company recognizes “the importance of neutrality in AI governance efforts,” and could be willing to be more open about how it is managing AI security and safety risks, he told Computerworld.

“While the need to innovate fast has strained governance for AI, increasing government scrutiny and the risk of public blowback is gradually bringing it back into focus,” Sengupta said. “It is likely that we will increasingly see independent third parties involved in AI governance and audit.”

AI to create better products and services, add $19.9T to global economy — IDC

Business spending to adopt and use AI in existing operations, and to deliver better products and services, is expected to drive 3.5% of global GDP by 2030, adding $19.9 trillion to the world economy, according to a new report from research firm IDC.

AI spending by businesses alone is expected to reach $632 billion by 2028, IDC had estimated in an earlier study.

As a result, AI will affect jobs across every region of the world, affecting industries from contact center operations to translation, accounting, and machinery inspection, according to IDC. Helping to trigger this shift are business leaders, 98% of whom view AI as a priority for their organizations.

David Foote, chief analyst and research officer with IT research firm Foote Partners, believes that 20% to 25% of tech jobs could eventually be taken by AI. “There have been a lot of layoffs,” he said. “Companies are identifying people who may have been solid workers in the past, but they don’t fit into the new world driven by the [emerging] economy and the technology they’re making bets on.”

While AI will reduce or eliminate the need for human input in some areas, it will also enhance productivity, requiring professionals to reskill and adapt to take on more strategic and creative roles, according to a research note by Foote. Along those lines, Goldman Sachs has projected that as many as 29% of computer-related job tasks could be automated by AI, as well as 28% of work by healthcare practitioners and technical tasks in that field. Careers with the highest exposure to AI automation are administrative positions (46%) and tasks in legal (44%) professions.

Nearly half of respondents to IDC’s Future of Work Employees Survey (48%) expect some parts of their work to be automated by AI and other tech over the next two years — and another 15% think most of their jobs will be automated. Only 3% expect their jobs to be fully automated.

Despite that disruption, however, AI will have a “net positive global economic impact,” according to the latest IDC report. In 2030, every new dollar spent on business-related AI solutions and services will generate $4.60 in the global economy in terms of indirect and induced effects.

Those impacts include:

  • Increased spending on AI solutions and services driven by accelerated AI adoption;
  • Economic stimulus among AI adopters, seeing benefits in terms of increased production and new revenue streams;
  • And an increase in revenue across the AI providers supply chain including services providers.

“In 2024, AI entered a phase of accelerated development and deployment defined by widespread integration that’s led to a surge in enterprise investments aimed at significantly optimizing operational costs and timelines,” said Lapo Fioretti, an IDC senior research analyst. “By automating routine tasks and unlocking new efficiencies, AI will have profound economic consequences, reshaping industries, creating new markets, and altering the competitive landscape.”

New job roles to emerge

survey of CFOs in June by Duke University and the Atlanta and Richmond Federal Reserve banks found that 32% of organizations plan to use AI in the next year to complete tasks once done by humans. And in the first six months of 2024, nearly 60% of companies (and 84% of large companies) said they had deployed software, equipment, or technology to automate tasks previously done by employees\.

While some work is being negatively affected by the rapid proliferation of AI tools and platforms, new positions such as AI ethics specialists and AI prompt engineers will emerge as dedicated roles within global organizations.

IDC’s research also indicates that positions where human social and emotional capabilities are critical, such as nursing and roles where decision-making encompasses ethics and comprehension beyond numbers, will remain robust. “Understandably, we’re all curious to know if AI will replace our jobs,” said Rick Villars, IDC group vice president, for worldwide research.

As one CEO told IDC researchers, “Based on this research it’s clear that we should be asking ourselves how our jobs can be made easier and better by AI. AI will not replace your job but someone who knows how to use AI better than you will,” Villars added.

How Apple’s quiet visionOS update hints at its plans

Has Apple’s love affair with AR cooled? Not on your life, and while it is true that visionOS didn’t get much time during the company’s big reveals at WWDC and last week’s iPhone launch, the company has delivered a handful of valuable improvements to the OS — hinting at future product plans as it did.

Perhaps you missed the hint, but what was interesting about the update wasn’t what’s new as much as what isn’t new, particularly the lack of support for Apple Intelligence. You could argue that this is only because the AI isn’t quite ready yet, but I can’t help but see its absence as a hint of what’s to come.

It’s always useful to cast about for a morsel of what Apple has actually said to support the theory. So, what has the company said? It’s already told us it’s working on a version of Siri with more contextual intelligence it hopes to ship in 2025.

That AI will be able to make contextual generative AI (genAI)-driven decisions in reaction to what it sees you interacting with on your iPhone, iPad, or Mac. Imagine what it will do if used with Vision Pro as it also looks around where you really are. If I’m right about this direction, it’s possible we’ll see visionOS equipped with a profoundly powerful contextual intelligence perhaps toward the end of next year

The Gift of Sound & Vision

Did I call this profound? Think about it: contextual intelligence is essential for an effective/responsive voice-driven interface at the intersection of technology and the everyday world. What you are looking at will change depending on your context, and the data your device surfaces will reflect the complexity of such complex lives. I see this as being Door Detection on steroids.I also think the gap between the idea and the reality will delay complete realization for a while — but it’s a beginning.

I also think that late 2025 time frame hints at Apple’s target release schedule for the slightly more mass market Vision 2.0 devices speculators expect. However, speculation doesn’t mean much these days until Bloomberg’s Mark Gurman’s “little birds” (Game of Thrones reference) begin to whisper. And I don’t think they’ve discussed how AI, voice, and contextual genAI will underpin dramatic new user interfaces across a multitude of consumer devices (plausibly including something designed by Jony Ive).

Let’s move away from speculation based on what didn’t happen to run through what’s actually changed.

Here’s a run-down on how Apple has expanded Spatial Reality:

The handy gesture

For me, the biggest improvement is around gesture. Apple has made it handy to access Home View and Control Center on Vision devices. To get to Home, you just need to use your hand — stare at your palm, then tap the dot that appears. If you turn your hand around you’ll be presented with time and battery information and can tap in that view to invoke Control Center or adjust volume controls. You can also now change the icon arrangement in Home, and avatar hand movements have been made smoother with new animations

Memories get Spatial

Apple made several tweaks to photos and videos: 

  • You can turn existing photos into spatial images, adding depth to create a stereoscopic effect.
  • The Photos app on Vision devices has been improved.
  • You can share photos, videos, and panoramas during FaceTime calls using SharePlay.
  • You get video trimming controls to use from within the headset.

(I’m quite interested to see the extent to which a future version of Apple Intelligence will be able to generate 3D environments from 2D photos you can then explore using Vision Pro. I believe that is inevitable.)

When you need a keyboard

You can use a Mac keyboard with Vision devices. In visionOS 2, the device will recognize your keyboard and display it on screen. This makes it much easier to use the input device. Apple has also introduced support for Bluetooth mice, which means you can navigate your device using a mouse and keyboard. A Messaging improvement means you can now dictate a message by staring at your microphone icon.

The infinite workspace

For work, perhaps one of the better enhancements (coming later this year) is the introduction of a new panoramic screen, equivalent to two 4K displays standing alongside each other. This really is giant real estate and should make complex workflows more possible.

On the web

The Safari browser lets Vision Pro users watch videos on a large display in any environment. Siri will read page content and Tab Group support makes it easier to handle multiple tabs. When you get time off, you get emoji reaction and singalong tools in Apple Music and the capacity to watch up to 5 MLS and MLB games in Multiview mode. (The latter feature is also expected later in the year.) You can also watch video in one window while working in other applications.

Virtuality and immersive environments

Apple expanded the number of immersive environments available in visionOS. It also improved the avatar system, so it captures more accurate skin tones and clothing colors. 

And the rest

  • Guest mode gives guests 30-day access to your device as it saves their eye and hand data.
  • Live Captions provide real‑time transcriptions of speech, audio, and video content, including FaceTime calls.
  • There’s a new travel mode for trains.

Please follow me on LinkedInMastodon, or join me in the AppleHolic’s bar & grill group on MeWe.