According to new research from Capgemini, Britain could actually be leading the AI adoption wave among European governments, with 75% of public sector organizations in the UK exploring or actively working with GenAI.
However, the same number of UK public sector orgs are also worried about generative AI's environmental impacts, with even more worried about data security and data sovereignty (78% each) and almost as many constrained by costs (68%).
Still, despite remaining apprehension, the UK is 11 percentage points ahead of the global average, where 64% of public sector organizations are exploring or using GenAI.
UK public sector is leading in GenAI adoptionAlthough the UK is currently ahead, it doesn't mean that the trend won't be challenged in the near future. As many as nine in 10 global public sector orgs plan to explore, pilot or implement agentic AI in the next two to three years.
With agentic AI marking the AI of tomorrow, Capgemini explored where today's AI, generative AI, is impacting most. Defense agencies (82%), healthcare (75%) and security sectors (70%) were among the highest adopters.
However, with such strict data privacy and security measures in place across the public sector, organizations are struggling to refine their own AI and are instead being forced to use off-the-shelf solutions. Only one-fifth (21%) say they have the necessary data to train and fine-tune AI models.
Moreover, just 12% feel very mature in activating data, and even fewer (7%) report maturity in data and AI skills. Capgemini Public Sector Global Industry Leader Marc Reinhardt summarized: "With rising citizen demands and stretched resources, public sector organizations recognize the ways in which AI can help them do more with less."
Looking ahead, 24% plan to appoint a Chief Data Officer and 41% plan to introduce a Chief AI Officer, on top of the 64% and 27% (respectively) that already have, bringing the number of public sector orgs with CDOs and CAIOs up to 88% and 68%.
Highlighting the importance of "the right data infrastructure," Reinhardt added: "Looking ahead, governments can be more agile and effective as AI augments the work of government employees to source information, conduct policy analysis, make decisions, and answer citizen queries."
You might also likeWith just a couple of months to go until the expected launch of the Samsung Galaxy Watch 8, some freshly leaked renders of the Classic version of the smartwatch have emerged online, showing the return of the device's rotating bezel.
These renders come from well-known tipster @OnLeaks and SammyGuru, and are based on schematic information leaked from the supply chain. We get to see the watch from all angles, and it looks like a stylish, polished gadget.
This particular tipster has a strong track record when it comes to revealing device designs ahead of time, but bear in mind that this is still just a leak – we won't know the final design of the Galaxy Watch 8 Classic until Samsung makes it official.
We didn't get a Classic model last year, but the upcoming watch looks to be combining elements from the Galaxy Watch 6 Classic from 2023 and the Galaxy Watch Ultra, which made its debut in 2024.
Rumored dimensionsIt seems we're getting a slightly squircle-shaped frame, as on the Galaxy Watch Ultra, though here it looks to be a little more rounded. It might actually be that the Galaxy Watch 8 Classic takes the place of the Samsung Galaxy Watch Ultra 2 this year.
Then there's that rotating bezel, which is always present on the Classic model, and which makes operating the watch easier. In recent years we've had Classic models every other year, so we're due another in 2025.
This same leak mentions a 1.5-inch display, and dimensions of "roughly" 46 x 46.5 x 14.2mm – which actually sound rather precise. The indications are that this will be the only size available when it comes to the Galaxy Watch 8 Classic.
If Samsung sticks to the same schedule it followed in 2024 we'll see this watch unveiled at an Unpacked launch event sometime in July, alongside the Galaxy Watch 8, the Galaxy Z Fold 7, the Galaxy Z Flip 7, and perhaps the tri-fold Samsung Galaxy G Fold.
You might also likeRicoh has announced the development of the Ricoh GR IV, the next generation of one of the best compact cameras. As a longtime GR IIIx user, I couldn't be more excited for what's in store.
We've been speculating for months if not years about when and if a GR IV would be unveiled. Along the way, leaks have given us false hope as further iterations of the GR III have been launched instead, such as the GR III HDF.
Now, finally, we know (roughly) when the GR IV will launch – a press release says we can expect the GR IV in "autumn 2025", while a niche highlight diffusion filter (HDF) version is set to follow in "winter 2025". There's no word yet on an 'x' iteration.
In the build-up to those launches the GR III will be discontinued, although production of the GR IIIx is set to continue for now.
I have already written about the upgrades I hope to see in the Ricoh GR IV, and now the press release has teased a few things that we can look forward to.
I loved the Ricoh GR IIIx concept so much that I personally forked out for one instead of upgrading to the latest iPhone, and I can't wait to see what improvements are made in the overdue next-gen model. (Image credit: Future)How big an upgrade will the Ricoh GR IV be?Ricoh says the "GR IV inherits the basic GR-series concept, while also incorporating totally new components, including a newly designed lens, a new image sensor and a new imaging engine, to further upgrade image quality.
"It also features more advanced communications functions and is compatible with a new application designed for smart devices to improve operability and functionality."
Those improvements could indeed make the GR IV, in Ricoh's words, "the ultimate snapshot camera in the history of the GR series"; but as an actual user of the GR IIIx there are some less headline-worthy upgrades that I'd also like to see.
Image quality from the GR III series is already stellar, but I hope we see some design improvements, particularly better build quality, a built-in flash and, provided the truly compact dimensions remain, a tilt LCD screen.
Performance-wise, the GR IV needs better autofocus, and the mention of a new processor and sensor gives me hope that'll be the case.
Ricoh says GR IV prototypes will be showcased at the GR Space showrooms in Tokyo, Beijing, and Shanghai from May 31. That should give us further clues as to any design changes, including the new lens.
The Ricoh GR IV is a camera I've been looking forward to for a long time, and with premium compacts particularly popular right now it can't come soon enough, and is set to be one of 2025's most popular cameras.
You might also likeSignal, a messaging app that majors in security and privacy (and has been in the limelight recently), has introduced a measure to defend content sent via the platform from the prying eyes of Microsoft’s Recall feature.
In case you forgot – unlikely, I know – Recall is the controversial feature for Copilot+ PCs that provides an AI-powered deep search in Windows 11. It does this by taking regular screen grabs of the activity on your PC, and you can see where the conflict comes in if the feature screenshots messages sent via a privacy-focused app like Signal.
As a counter to this, Signal announced in a blog post (picked up by The Verge) that it’s introducing a new ‘screen security’ setting for its Windows 11 app, and this will be turned on by default for users of the desktop OS. This functionality is rolling out now.
What the developer has done here is activate a DRM flag on the Signal app window as a fudge to stop Recall from capturing screenshots. It’s using that workaround because Microsoft failed to offer “granular settings for app developers that would enable Signal to easily protect privacy” with Recall, the blog post observes.
Signal notes: “We are enabling an extra layer of protection by default on Windows 11 in order to help maintain the security of Signal Desktop on that platform even though it introduces some usability trade-offs. Microsoft has simply given us no other option.”
What’s the usability trade-off? Blocking screenshots from every angle using this DRM fudge means legitimate use of screen grabbing is ruled out, and also some screen reading tools (like Narrator in Windows 11) or other accessibility features may not work properly. So, that’s far from ideal, but Signal argues that Microsoft has left it no choice.
Analysis: Clear Signal(Image credit: Signal)This does seem to be an important part of the equation that’s missing with Recall. Famously, Microsoft pulled the feature after it was first revealed a year ago, and took it back to the drawing board, tightening up security and privacy in numerous ways. But that was a case of fixing and smoothing over weaknesses, rather than strengthening measures – Recall was launched way too early, and without enough thought, worryingly.
And still after all that time, there are software developers out there like the maker of Signal pointing out a basic flaw in Recall that Microsoft should have addressed at this point.
Remember that this doesn’t apply to you if you don’t have a Copilot+ PC, though, as only those devices get Recall (it requires a beefy NPU for locally accelerating AI workloads to ensure the feature works smoothly enough).
While the new setting is enabled by default for Signal in Windows 11, you can turn it off by going to Signal Settings > Privacy > Screen security (you’ll be warned that you’re about to disable the ability).
It’ll certainly be worth watching this space going forward, and I’m hoping Microsoft will implement a proper way to let developers control how Recall interacts with their apps, as this only makes sense – particularly for privacy-related software.
You might also like...Amid ongoing US-China tensions over tech dominance, Google has recognized that some countries are concerned about the US' tech dominance, thus it has announced some updates to its sovereign cloud services to ensure optimal privacy.
In a blog post, Google Cloud President of Customer Experience Hayete Gallot boasted that the company now has more than 42 cloud regions, 127 zones, 202 network edge locations and 33 subsea cable investments supporting its global cloud services.
Once niche and limited to highly regulated sectors like defense and intelligence, sovereignty is now a mainstream concern for many sectors and most governments, and Google has responded with a handful of updates.
Google upgrades its sovereign cloud portfolio for customersKey to the announcement is Google Cloud's air-gapped cloud service, which operates as a standalone cloud with no external network connectivity requirements. It's designed specially for sectors with strict data residency and security, and is build on open-source components which Google believes can enhance resilience and therefore business continuity.
The company even got authorization to store US government Top Secret and Secret-level data on Google Cloud Air-Gapped in 2024 – a sign of its strong privacy credentials.
The second of Google's three key announcements, Google Cloud Dedicated, is a partner-operated regional platform that meets local sovereignty standards. The company already partnered with Thales in 2021 to make its first-ever Trusted Cloud by S3NS for Europe, but now it's confirmed it'll be targeting Germany next.
Finally, Google Cloud Data Boundary gives users further control over where their data is stored and processed, with enhanced protection via external key management and confidential computing. For example, Workspace users can select to restrict processing to the US or EU, while also choosing local data storage countries.
A Mandiant-backed tool called User Data Shield is being added to this to ensure the ongoing verification of sovereignty postures.
Gallot expressed a commitment to "empowering organizations globally to navigate the complex landscape of digital sovereignty with confidence," while also noting how advanced security measure like a zero-trust posture, post-quantum cryptography and AI-powered defenses form part of the broader strategy.
You might also likeIf you’re still rocking an iPhone 7 Plus or an iPhone 8, we’ve got some bad news: Apple has just added them to its list of “vintage” products, meaning your repair options have just become a little more limited.
Apple declares a product to be vintage when the company stopped selling it more than five but less than seven years ago. Apple does this from time to time with older devices – after all, it can’t offer ongoing support for every device it's ever produced on an indefinite basis. But the news will still come as a disappointment to anyone still using one of these phones.
What does this mean in practice? Well, when Apple classifies a product as 'vintage', it cuts back on the support it will offer to owners. In theory, you’ll still be able to get one of these products repaired by Apple or an Apple Authorized Service Provider (AASP), but it depends on whether the parts are still available. If they aren’t, you won’t be able to get genuine components when it’s time to repair your phone.
After a product has been classed as vintage, the next step is for Apple to label it 'obsolete', and this usually occurs when more than seven years have passed since Apple last offered the device for sale. The iPhone 7 Plus and iPhone 8 were discontinued in 2019 and 2020, respectively, which means you’ve still got another year or two until they're classed as obsolete.
Old-school charm(Image credit: Future)Launched in 2016, the iPhone 7 Plus was the larger version of the iPhone 7, and it came with water and dust resistance, a solid-state Home button, new color options, and revised antenna bands. It also controversially marked the removal of the 3.5mm headphone jack once and for all.
The iPhone 8, meanwhile, was released in 2017 and came with wireless charging, True Tone technology, a better chip, and improved cameras. It launched alongside the iPhone X, which was the first all-screen iPhone and consigned the Home button to history.
This is the second time in recent months that Apple has added devices to its vintage products list. In mid-April, the iPhone 6S was marked as vintage, alongside the 2018 Mac mini. Don’t be surprised to see further products added to Apple’s list over the next year.
You might also likeSony has announced that it's shutting down its PlayStation Stars loyalty program after three years.
In a new PlayStation Blog post, the company confirmed that PlayStation Stars will be closing down and will no longer accept new members as of today.
Those who choose to cancel their membership won't be able to rejoin and will lose all earned Points.
"Since launching the program, we’ve learned a lot from evaluating the types of activities our players respond best to, and as a company, we are always evolving with player and industry trends," Sony said.
"Through this evaluation, we have decided to refocus our efforts and will be winding down the current version of PlayStation Stars. We will continue to evaluate our key findings from this program, and are looking into ways to build upon these learnings."
Although the rewards program is shutting down, members can still earn Digital Collectibles, Points, and level up their status until July 23, 2025 at 7:59AM PDT / 3:59PM BST / 11:59PM JST.
After this date, no new PlayStation Stars Campaigns will be available, or additional rewards or status benefits to earn.
Sony said that "this version" of PlayStation Stars will fully end on November 2, 2026, but it's unclear if the company is planning on releasing another version of the loyalty program.
It also said that current PlayStation Stars members can continue to redeem Points as long as their Points don’t expire before that date, and that members can still access their Digital Collectibles, even after these changes go into effect.
The program was made available for PS5 and PS4 in 2022 and rewards players with earnable loyalty points that can be redeemed for various digital goodies after completing various challenges, like unlocking a specific trophy or playing a game on your system.
You might also like...New research from ICANN has revealed that generic top-level domains (gTLDs) have been likened to digital storefronts, noting that unique online addresses are becoming critical for branding and customer engagement.
ICANN's global survey of more than 2,000 marketers from 14 global markets found that gTLDs like .microsoft or .brazil offer untapped branding and operational opportunities.
The study also found that brands can enhance their identity, improve trust and control their online presence more effectively when using a gTLD.
Businesses are missing out on generic TLD benefitsAn overwhelming majority (92%) of the marketers surveyed agreed that gTLDs offer benefits, including enhanced brand differentiation/identity (46%), improved customer trust and engagement (45%), better control over online presence (44%) and improved SEO (44%).
Although gTLDs might be new additions to companies' roadmaps, they can also form part of existing strategies. For example, 53% noted wanting to stand out from competitors as a top priority in the next 12 months. One in two (52%) also noted wanting to attract and engage the right audience, while a similar number (47%) declared keeping up with digital trends to be a top priority.
However, as with any change in technology, some are finding that there are certain hurdles preventing them from going all-in on gTLDs, including high costs (31%), a lack of knowledge (27%) and insufficient time, unclear ROI or security concerns (24% each).
ICANN also observed misconceptions, with 16% of marketers believing gTLDs were government-only and 9% thinking they were part of trade agreements.
With gTLDs offering a transformational opportunity for brands willing to spend the time setting them up, ICANN is urging brands to start preparing now in order to arm themselves with the knowledge and power.
You might also likeXbox has partnered with classic game service Antstream Arcade to bring Xbox Game Pass subscribers the new Retro Classics collection.
Microsoft announced the partnership via an official Xbox Wire post, which is described as "a collaboration between Xbox and Antstream Arcade, available to play for Game Pass members globally."
The service is available now and only requires the standard Xbox Game Pass subscription - Ultimate is not necessary. The Retro Classics collection is a separate app, available on Xbox Series X|S, PC, and can also be streamed via supported Samsung and LG smart TVs, Amazon Fire TV devices and Meta Quest VR headsets.
Initially, the Retro Classics library houses over 50 retro Activision games. These include Pitfall, Dragster, Grand Prix, Kaboom!, Enduro and many more. Games also have their own sets of achievements, leaderboard challenges, and save state support - not unlike the main Antstream Arcade service.
"This initiative is a step in our commitment to game preservation and backwards compatibility, allowing players to experience many timeless games on modern devices," the Xbox Wire post reads.
Furthermore, the Retro Classics collection is set to expand over time. The post claims that the launch "is just the beginning." Eventually, the service is set to grow to well over 100 playable titles. If you don't have the cash for some of the best retro games consoles, then this could be a more affordable solution.
It's a pretty smart way to leverage a suite of classic games now made available to Microsoft via its Activision Blizzard buyout, and one I didn't really see coming. The Antstream Arcade service itself is something I've personally gotten some enjoyment out of, and it's cool to see a similar service make its way to Xbox Game Pass members on both console and PC.
You might also like...The Samsung Galaxy S25 has the same main camera sensor as the Samsung Galaxy S24, so while this camera performs reasonably well, it’s really due an upgrade with the Samsung Galaxy S26. Thankfully, an upgrade is what we might get.
According to GalaxyClub, the Samsung Galaxy S26 will have a new 50MP sensor. Now, that’s no change in megapixels, but the site claims that “as far as we understand” it will be a new model in Samsung’s ISOCELL GN series, and possibly one that hasn’t been announced yet. That would be in place of the 50MP ISOCELL GN3 in the Samsung Galaxy S25.
While there’s no mention here of the Samsung Galaxy S26 Plus, we’d expect that would also have this new sensor since the Samsung Galaxy S25 Plus has the same sensor as the base S25. That’s assuming there is a Galaxy S26 Plus at all, as some reports claim Samsung is considering ditching this model in favor of a Samsung Galaxy S26 Edge.
Samsung or Sony?Speaking of the S26 Edge, leaker @Jukanlosreve has claimed that both this and the base Samsung Galaxy S26 will use a 50MP Sony sensor. So that’s slightly at odds with the claim above, as while both leaks point to a new 50MP sensor for the Galaxy S26, there’s disagreement on the brand.
BIG: * NEXT YEAR'S S26 VANILLA AND EDGE MODELS WILL FEATURE A 50MP 1.0UM SENSOR FROM SONY.May 22, 2025
Still, it seems likely then that the sensor will be changing, even if we’re not sure what to. Equipping the Samsung Galaxy S26 Edge with a 50MP sensor would seem an odd choice though, since the Samsung Galaxy S25 Edge has a 200MP one. Fewer megapixels doesn’t necessarily mean a reduction in quality, but it might not look great from a marketing perspective.
We’d take all of this with a pinch of salt anyway, as the Samsung Galaxy S26 series probably won’t land until early 2026.
We have started hearing quite a few things about these phones though, with earlier leaks pointing to bigger batteries across the whole S26 series, while the base model could have an Exynos 2600 chipset in some regions, and the Samsung Galaxy S26 Ultra could get a new lens and a variable aperture.
So it sounds like there could be quite a lot of upgrades coming to Samsung’s flagship phone series.
You might also likeSpeed is no longer a luxury for businesses, it's a necessity. Whether shopping online, streaming content or using business apps, users expect instant and seamless experiences. Simultaneously, companies face increasing pressures to innovate, scale, and remain secure against ever-intelligent threats, all without breaking the bank.
As the internet and technology stacks evolve and grow increasingly complex, legacy IT infrastructures have struggled to keep up, resulting in poor user experience, security vulnerabilities and runaway spending. Many businesses are now turning to a modern edge architecture that seamlessly integrates in their stack and way of working, allowing for faster innovation, easier pivots and instant deployments. Here are four reasons why modern production systems need an edge architecture to match.
Being slow is unacceptableToday’s audiences expect engaging and immersive content, and are more impatient than ever with low tolerance for low quality resolution. Viewers expect content to play instantly, every time.
Adding to the challenge of making your experiences instant every time, content is rarely consumed where it’s created, and geographical distances will introduce unavoidable latency. The more popular you are, the broader your geographical reach and the more global demand for instant playback.
Without a proper architectural strategy, a system broadcasting breaking news, viral content, or a game-winning moment in the United Kingdom will have Australian viewers waiting for data – an issue compounded by network congestion, routing inefficiencies, and the occasional ship’s anchor.
While the issue is as old as the Internet, edge computing solves the inherent delay using points of presence (POPs) placed strategically across the globe. At any given time, POPs can cache and deliver content from the server nearest the consumer, ensuring they receive content with the smallest possible delay. Latency is drastically reduced, page load times are accelerated, and consumers will experience improved performance for applications, streaming services, and websites.
Success shouldn’t break your budgetOn the surface, using two or more hyperscaler zones as a way to build a geographically distributed application seems straightforward – and this is valuable for disaster recovery purposes. However, to do this in a way that simulates a true edge system with dozens of locations is not feasible when you consider cloud costs and engineering resources. Moreover, it distracts from your innovation efforts with your core business.
Modern distributed edge systems are more capable, flexible, and affordable than ever. They provide proven, robust performance with global coverage, enabling fast, personalized, and secure computing. Equipped to give developers the tools they need for instant scale and custom computing, these systems have evolved from traditional CDN use cases into highly customizable, API-friendly platforms that can integrate with a wide range of demanding applications. And you won’t need to spend tens of millions on physical infrastructure and operational talent to leverage them.
By using distributed datacenters, content owners and online businesses can optimize performance without extensive hyperscaler investments. Edge systems also help reduce cloud bandwidth costs by caching frequently accessed content closer to end users, minimizing egress traffic. This not only lowers operational expenses but also ensures an infrastructure that can deliver at scale when your business has the most successful days of its existence.
Your users require privacy and securityUsers today are savvier than ever regarding their data privacy and security. With a new paradigm of regulatory requirements like GDPR, businesses must be prepared to simultaneously meet these requirements while also delivering the experience their customers expect within budget.
It is essential for companies that deal with sensitive data to leverage modern architectural design to separate public and private information without interrupting the user experience. One possible way to do this is to organize the data into separate API responses or services into static, cacheable data (financial market trends, news, product information) and personalized, non-cacheable data (account balances, user-specific notifications).
This way, public data can be cached while private data is kept secure and private. It also conforms to the trends in mobile app development, where API-first architectures naturally separate user-specific calls from general data. All of this is instantly configurable and programmable with a modern edge partner.
Beyond privacy expectations, for industries facing piracy threats and account takeover risks, edge computing systems offer distributed and high-performance countermeasures. A modern threat model requires a security portfolio that is intelligent, adaptable and cost-effective. An edge-based solution localizes enforcement so that platforms can detect and prevent unauthorized access in real-time. And since performance is optimized, users benefit.
Decoupling data, taking advantage of edge programmability, and incorporating dynamic caching mechanisms allow businesses to increase security, improve performance, and lighten the load on their infrastructure without affecting the user experience. Your architecture strategy should tightly integrate each of these perspectives.
Focus on reliabilityEven if your site or applications are fast, within budget, and manage your user’s privacy and security perfectly, it’s all for nothing if your uptime flakes or your team is constantly paged out of bed to fight midnight incidents. Thankfully, the same edge strategies can give your business layers of resiliency through intelligent caching, layered origin protection, and instant system scalability.
Don’t let the biggest and most successful days for your business turn into catastrophes due to a lack of architectural planning and the wrong edge partner. Plan ahead and choose the most flexible, cost-effective, secure, and resilient partner to plan for your long-term success.
We've compiled a list of the best business smartphones.
This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
In March, AWS announced the general availability of its new multi-agent capabilities, bringing the technology into the hands of businesses across almost every industry. Until now, organizations have mostly relied on single-agent AI systems, which handle individual tasks but often struggle with complex workflows.
These systems can also break down when businesses encounter unexpected scenarios outside their traditional data pipelines. Google also recently announced ADK (Agent Development Kit) for developing multi-agent systems and A2A (Agent to Agent) protocol for agents to communicate with each other, signaling a broader industry shift toward collaborative AI frameworks.
The general availability of multi-agent systems changes the game for startups. Instead of a single AI managing tasks in isolation, these systems feature robust and manageable networks of independent agents working collaboratively to divide skills, optimize workflows and adapt to shifting challenges. Unlike single-agent models, multi-agent systems operate with a division of labor, assigning specialized roles to each agent for greater efficiency.
They can process dynamic and unseen scenarios without requiring pre-coded instructions, and since the systems exist in software, they can be easily developed and continuously improved.
Let's explore how startups can leverage multi-agent systems and ensure seamless integration alongside human teams.
Unlocking value for startupsStartups can leverage multi-agent systems across several critical business functions, beginning with research and analysis. These systems excel at data gathering, web searches, and report generation through the process of retrieving, organizing and dynamically refining information.
This allows systems to streamline complex research workflows, enabling startups to operate more efficiently and make informed decisions at scale. Meanwhile, in sales processes, multi-agent systems improve efficiency by automating lead qualification, outreach and follow-ups. AI-driven sales development representatives (AI SDRs) can automate these repetitive tasks, reducing the need for manual intervention while enabling teams to focus on strategic engagement.
Many startups may also need to extract structured data from unstructured sources. For example, multi-agent systems automate web scraping and adjust to website format changes in real time, eliminating the need for continuous manual maintenance.
Unlike traditional data pipelines that require constant debugging, multi-agent systems autonomously manage tasks, reducing the need for large development teams. This is particularly useful for startups as they can ensure up-to-date data without expanding technical teams too quickly.
How businesses can implement multi-agent systemsStartups seeking to gain outsized results by leveraging these systems can do so through two impactful approaches.
One option is purchasing existing solutions to replace complex data flows and human-driven processes. This is the most cost-effective choice for many startups, as they can automate and replace complex sales pipelines and make data workflows more robust, reducing reliance on humans for repetitive tasks.
But for startups with unique operational needs, developing a multi-agent system in-house is ideal. Traditional systems require coding for every possible scenario – a rigid and time-consuming approach that is prone to human error. Multi-agent systems, in contrast, are tailored for all possible scenarios and dynamically adapt to complexities, making them a more flexible and scalable alternative.
Regardless of whether startups buy or build, multi-agent systems provide a game-changing opportunity to streamline operations, reduce manual workloads and improve scalability.
Overcoming challenges in AI integrationDespite its advantages, integrating multi-agent systems comes with certain challenges. Decision-making by agents within the multi-agent system isn’t always transparent since the systems often rely on large language models (LLMs) that have billions of parameters. This makes it challenging to diagnose failures, especially when a system works in one case but fails in another.
Additionally, multi-agent systems deal with dynamic, unstructured data, meaning they must validate AI-generated outputs across various input sources - from websites to documents, scanned documents and chat and meeting transcripts. This makes it a greater challenge to balance robustness to changes and accuracy. Beyond this, multi-agent systems face difficulties in maintaining effectiveness and require monitoring and updates in response to input source changes, which often break traditional scraping methods.
Startups can overcome these challenges by embracing new tools, such as LangFuse, LangSmith, HoneyHive and Phoenix, which are designed to enhance monitoring, debugging, and testing in multi-agent environments. Equally important is fostering a workplace culture that embraces AI agents as collaborators, not replacements. Startups should ensure buy-in across stakeholders and educate employees on the value of AI augmentation to allow a smooth adoption.
Transparency is also key. Founders must be open with staff about how multi-agent systems will be used to ensure a smooth collaboration between human and AI coworkers.
Achieving outsized resultsThe AI field is moving fast, making it difficult for experts, let alone everyday users, to keep up to date with each new model or tool that is released. Some small teams may therefore see multi-agent systems as unattainable.
However, the startups that successfully implement them into their workstreams – whether by purchasing or building custom solutions – will gain a competitive edge. Multi-agent systems bridge the gap between AI and human collaboration that can’t be achieved with traditional single-agent systems.
For startups focused on growth, multi-agent systems are the best tool in their arsenal to compete with incumbents who might be stuck with an outdated tech stack. The ability to streamline operations, reduce manual workload, and scale intelligently makes multi-agent systems an invaluable tool in achieving outsized results.
We've compiled a list of the best landing page creators.
This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
Fujifilm has outdone itself with the new X half – a retro compact camera that packs some of its wackiest and outright funnest ideas yet, all inspired by film photography.
There’s a clue to the X half’s inspiration in the name – it’s a digital reimagining of half-frame film cameras like the Pentax 17. I've already tried the X half, and it was a much needed dose of fun – check out my X half hands-on review.
To facilitate half frame, the X half's 18MP JPEG photos are taken in 3 x 4 vertical format, recorded onto a vertical 1-inch sensor, and composed using the unique vertical LCD.
Alongside that fixed screen is a secondary screen that mimics the film canister window you see on many film cameras, and there's a fun surprise here – it’s touch sensitive, and allows you to swipe up or down to select one of Fujifilm’s Film Simulations. If this charming feature doesn't make its way into future Fujifilm cameras, I'd be shocked.
Film Simulation color effects are well known – they're inspired by Fujifilm film stock, and have helped to cement Fujifilm’s popularity over the last 10 years through cameras like the X100VI. The X half offers a stripped-back selection of 13 popular Film Simulations, including Provia and Astia.
You'd think all of the above would be enough to secure the X half's unique status, but Fujifilm has really let loose, with even more features for film photography fans to enjoy.
The LCD emulating a film cannister window with Velvia Film Simulation, and the vertical LCD (Image credit: Tim Coleman)Simulating film to another levelGoing one step further from that twin-screen combo and vertical shooting, there’s a Film Camera mode. This locks in your chosen Film Simulation and camera settings such as ISO, and disables the screen preview, leaving you to compose your shots via the optical viewfinder instead, as if you're shooting with film.
Once your ‘film’ is used up – either 36, 54 or 72 shots – you can exit the mode and view the screen once more, and make changes to settings again.
Film Camera mode is such a fun feature, and for me is the closest experience to film photography that I've had using a digital camera – and it's optional.
The film wind lever tucked in with the camera off. In the on position, the lever sticks out for an easy reach. (Image credit: Tim Coleman)Then there's what is in effect a film wind lever, which in this case, through 'cranking', is used to create diptychs – that’s two vertical shots side by side. These are recorded individually through the vertical 1-inch sensor, but then composited afterwards and displayed just like you'd get with a half-frame film camera on a roll of 35mm film.
Again, you can take or leave the diptych feature. I reckon it's a nice to have – working out how image pairs complement each other stretches your creative muscles.
We also get some completely new picture effects, almost all of which are film photography-inspired and include light leak, expired film and halation.
Full HD video capture is also possible, and the diptych effect can be applied to both photos and videos, which is really neat.
This is all packaged in a palm-sized, premium-feel compact that features a fixed 32mm f/2.8 lens with a mechanical aperture, plus the same battery as used in cameras like the X100VI for an 880-shot life, and which weighs just 240g.
Fujifilm X half in charcoal (left), silver (middle) and black (right). (Image credit: Tim Coleman)Fujifilm has created a dedicated app for the X half, which can be used to make diptychs, and upload and view images, plus the camera can connect wirelessly to one of Fujifilm’s Instax printers for on-the-go printing.
The app wasn't available when I tested the camera, but will be downloadable from early June. Meanwhile, the Fujifilm X half itself will be available globally from June 12 in silver, charcoal and black, and costs $849 / £699 / AU$1,349.
I’ve been reviewing digital cameras for 15 years, and the Fujifilm X half has to be one of the funnest yet – a compact camera with a difference. You can configure it in a way that’s as close to a film camera as you’re going to get with digital, plus it packs the retro look and feel that we’ve come to expect from Fujifilm.
What do you think of the Fujifilm X half? Let us know in the comments below.
You might also likeI write about vacuum cleaners for a living, and while performance varies, most new models these days tend to look roughly the same.
So when news of a new addition to the Dyson vacuum lineup landed in my inbox, I expected to see something similar to its existing models: slick and high-quality, but not especially distinctive or surprising.
How wrong I was.
The newly unveiled Dyson PencilVac doesn't just have an unusual name, it's all-round one of the most unique vacuums I've seen. This brand knows what it's doing in this marketplace – it makes some of the best cordless vacuums you can buy, and today's very best Dyson vacuums include features you still can't find anywhere else.
So while the PencilVac strays a long way from the tried-and-tested formula of what works for vacuum cleaners, I'm very optimistic about its performance. Here's a rundown of the most intriguing features in this new launch...
1. It's ridiculously thinThe most immediately noticeable thing about the PencilVac is that it's incredibly streamlined. Without the floorhead, the whole thing is 1.5 inches / 3.8cm in diameter. To make that possible, the brand had to develop a tiny new motor – the Dyson Hyperdymium 140k motor is just 1.1 inches / 2.8cm wide, and hidden entirely within the handle.
The PencilVac is also impressively lightweight, clocking in at 4lbs / 1.8kg. For context, the lightest option in our best cordless vacuum roundup right now is 5.7lbs / 2.6kg, and there are a number of models that weigh over 6.6lbs / 3kg.
All the PencilVac's mechanics are shrunk down and fitted inside the handle (Image credit: Dyson)Generally, when you shrink down a vacuum, you sacrifice power. That's why handheld vacuums tend to be much less 'sucky' than full-sized options. That holds true for the PencilVac – there's 55AW of suction, compared to 115AW for the V8 (the oldest Dyson stick vacuum in the current range) and a massive 280AW for the latest-and-greatest Gen5detect. However, while it's unlikely to be suitable for a truly deep clean, that's still a decent amount of suction for the size and weight.
As a side note, the 1.5-inch / 3.8cm diameter isn't incidental. Brand founder James Dyson says, "I have long wanted to make a vacuum of only 38mm diameter (the same as my latest hair dryer, the Supersonic r)". The Dyson Supersonic r is the pipe-shaped dryer that was originally released for professionals only, but recently joined the main consumer range.
2. There are cones instead of rollersMoving down to the business end, and you'll find the new 'Fluffycones' floorhead. It sounds like a Pokémon, but it's actually a reimagined cleaner head. Vacuums traditionally have one brush roll, maximum two, and they're tube-shaped. The Dyson PencilVac has four brushrolls, and they're all conical.
There's logic to the tapering shape: it helps direct long hair along the roll and into the dust cup, whereas with parallel rollers the hair tends to just wrap around and stay there, until your rip it off or attack it with scissors. Dyson's hair screw tool also has a conical brush roll, and works exactly as it's meant to when it comes to tackling long hair.
Rather than one parallel brushroll, the PencilVac has four tapering rollers (Image credit: Dyson)The cones project out at the sides so they can clean right to the edges of rooms, and the whole thing can lie flat to the ground, with a clearance of just 9.5cm / 3.75 inches off the floor.
I'm interested in Dyson's description of the rollers as 'fluffy', because in the brand's vocabulary that usually indicates a soft roller for use on hard floors only. In fact, the more I look at this vacuum, the more I'm convinced it's a specialist model just for use on hard floor. It's not specified in the press material I have so far, but it would make sense with the lower suction and smaller dust capacity.
3. There's no visible dust cupOne of the most baffling things about the PencilVac is that it doesn't appear to have a dust cup. Of course, there is one – like the motor, it's hidden away inside the handle.
The capacity is next-to-nothing: just 0.08L. However, Dyson has introduced a dust compression system, which uses air to squish down the particles so they take up as little room as possible. Dyson claims that means it can hold five times the physical volume.
The dust cup is also hidden within the handle (Image credit: Dyson)The emptying process has also been reimagined, with a push-lever system replaced by an exciting-sounding "syringe, no-touch bin ejection mechanism".
As it pushes out dust and debris, the mechanism simultaneously wipes the 'shroud'. I'm not totally clear what the 'shroud' is in this context, but I do know that keeping the internal mechanisms clean is key to efficient vacuum performance, so this seems like a good thing.
4. The floorhead glows and appears to floatAs well as siphoning off hair as you clean, the floorhead cones have another trick up their sleeve. The cones rotate in opposite directions, the aim being that this vacuum cleans just as well when it's pushed forward as when it's pulled back. This is a bit of a weak spot on the regular Fluffy floorhead – it has no trouble sucking things up when moving forwards, but pull it back and debris will pool behind it.
I'm intrigued to see how this new approach works in practice – especially because Dyson describes it as "floating" across the floor. I wonder, too, if it might make this vacuum reversible altogether, given the fact that the handle section looks very symmetrical.
(Image credit: Dyson)Dyson has also added "laser-like" illumination to both the front and back of the floorhead. This is another feature borrowed from the exsiting Fluffy floorhead, and helps create big shadows on the tiniest bits of dust, which otherwise might go missed. It only works on hard floors, which is another indication this vac is likely not for carpet.
5. There's a tool that looks like a chimney brushThere's an intriguing addition to the tool lineup in the form of a 'Rotating combi-crevice tool', designed for cleaning in awkward gaps. This seems especially geared towards cleaning high-up, where it can be tricky to get your angles correct. It makes particular sense for an ultra-light vacuum like this one, which is far easier to lift above your head than your average stick vacuum.
As an aside, it looks like the PencilVac is button- rather than trigger-operated. That's dictated by the streamlined shape, but it's also great news for maneuverability and easy of use – the fact that many Dyson vacs still use a trigger to turn on is a perpetual bugbear of mine.
You'll also get a Conical hair screw tool, similar to the one included with the newest Dyson stick vacuums, for tackling long hair on furniture. Both can be stored on the magnetic charging dock.
The Rotating combi-crevice tool looks perfect for cleaning up high (Image credit: Dyson )6. It's app-connectedI'm much less excited about this feature, but it feel like I should point out that this is the first Dyson cordless vacuum to connect to the MyDyson app. The app will provide more information about battery life and also report on filter status. However, there's also a screen on the vacuum itself showing remaining battery, so I'm hoping the app connection is an optional extra rather than an essential.
There's a companion app, but key information is also shown on the vac's screen (Image credit: Dyson)Price & availabilityThe PencilVac will arrive in Australia first, with launch scheduled for August 2025. It's due to go on sale in the UK sometime in 2026, and I'm awaiting info as to if/when it will come to the US. As of yet I don't have any pricing info at all – I'll update this article with more details when I have them.
You might also like...Google has rolled out a new AI-powered shopping feature to help you figure out what the clothes you are interested in buying might look like when you wear them. It's dubbed "try it on" and it's available right now in the US through Google Search Labs.
To get started, you just need to switch it on in the lab. Then, you upload a full-length photo of yourself and start looking for clothes in the Google Shopping tab.
When you click on an image of some outfit from the search results, you'll see a little "try it on" button in the middle of the enlarged version of the outfit in the right-hand panel. One click and about ten seconds later, you'll see yourself wearing the outfit. It may not always be a perfect illusion, but you'll at least get a sense of what it would look like on you.
Google claims the whole thing runs on a model trained to see the relationship between your body and clothing. The AI can, therefore, realistically drape, stretch, and bunch material across a variety of body types.
The feature doesn't work with every piece of clothing you might see, or even every type of outfit. The clothing retailer has to opt into the program, and Google said it only works for shirts, pants, dresses, and skirts.
I did notice that costumes and swimwear both had no usable images, but I could put shorts on myself, and costumes that looked enough like regular clothes were usable. The AI also didn't seem to have an issue with jackets and coats as categories.
Elvis looks(Image credit: Photo/Google AI)For instance, on Google Shopping, I found replicas of the outfits Elvis wore for his 1966 comeback and one of his jumpsuits from the 1970s. With a couple of clicks, I could imagine myself dressed as the King in different eras.
It even changed my shoes in the all-black suit. I'd always wondered if I could pull off either look. The images are shareable, and you can save or send them to others from the Google mobile app and see how much of an Elvis your friends think you are.
Super summer(Image credit: Photo/Google AI)The details that the AI changes to make the photos work are impressive. I used the AI to try on a fun summer look and the closest to a superhero costume I could try. The original photo is me in a suit and jacket with a bowtie and black dress shoes. But the shoes and socks on both AI-generated images not only match what was in the search result, but they're shaped to my stance and size.
Plus, despite wearing long sleeves and pants, the AI found a way to show some of my arms and legs. The color matches reality, but its imperfections are noticeable to me. My legs look too skinny in both, like the AI thinks I skipped leg day, and my legs in the shorts have not been that hairless since I turned 13.
Imperfections aside, it does feel like this will be a major part of the next era of e-commerce. The awkward guessing of whether a color or cut works for your skin tone and build might be easier to resolve.
I wouldn't say it can make up for trying them on in real life, especially when it comes to sizing and comfort, but as a digital version of holding an outfit up against you while you look in a mirror, it's pretty good.
Ending unnecessary returns(Image credit: Photo/Google AI)Uncanny as some of the resulting images are, I think this will be a popular feature for Google Shopping. I'd expect it to be heavily imitated by rivals in AI development and online retail, where it isn't already.
I particularly like how the AI lets you see how you'd look in more outlandish or bold looks you might hesitate to try on at a store. For example, the paisley jacket and striped pants on the left or the swallowtail jacket and waistcoat with Victorian trousers on the right. I'd hesitate to order either look and would almost certainly plan on returning one or both of them even before they arrive.
Returns are a plague on online retailers and waste tons of packaging and other resources. But if Google shows us how we’d look in clothes before we buy them, it could chip away at return rates; retailers will race to sign up for the program.
It could also open the door to more personalized style advice from AI. You could soon have an AI personal dresser, ready to give you a virtual fit check and suggest your next look, even if it isn't something Elvis would have worn.
You might also likeNvidia CEO Jensen Huang has once again looked to highlight the huge potential AI can offer companies of all sizes in the coming months and years.
Speaking at the recent Dell Technologies World 2025 event, Huang noted “AI is here - this is unquestionably the single biggest platform shift.”
In conversation with Dell Technologies CEO Michael Dell, Huang added how, “from a technology perspective…we’re now in perception to generative to now reasoning AI models, and that’s at the raw technology level.”
"The biggest reinvention"Huang highlighted how Nvidia and Dell are teaming for enterprise AI, which he called, “one of the largest opportunities ahead of us”.
“These are companies that are essentially building a digital workforce of AI agents, which can be working in cybersecurity, software engineering, marketing and sales operations, and forecasting, and supply chain management - all these different AI agents are being created now, that can augment our human workforce with a digital workforce.”
(Image credit: Dell Technologies)One of the biggest announcements at Dell Technologies World 2025 concerned the expansion of Dell’s AI Factory platform, which has received some significant updates thanks to Nvidia.
Initially launched at DTW 2024, the next iteration of the Dell AI Factory, unsurprisingly called Dell AI Factory with Nvidia 2.0, encompasses client devices, servers, storage, data protection and networking
The new iteration includes six new servers, including the air-cooled PowerEdge XE9780 and XE9785, and the liquid-cooled XE9780L and XE9785L, all of which support up to 192 Nvidia Blackwell Ultra GPUs with direct to chip cooling.
These new releases can also be customized with up to 256 Nvidia Blackwell Ultra GPUs per Dell IR7000 rack, which Dell claims can deliver up to four-times faster large language model training than its predecessor.
The two companies also announced Dell Managed Services for the Dell AI Factory with Nvidia, which looks to simplify AI operations with the management of the full Nvidia AI stack, and 100-times faster token generation per second for distributed AI inferencing, with more than 80% reduction in latency, to help support the growth of agentic AI.
(Image credit: Future / Mike Moore)Noting that he and Huang had known each other “for some 30 plus years”, Dell asked the Nvidia CEO if he wanted to give any advice to the Dell Technologies World audience.
"This is a once in a lifetime opportunity - in the last 60 years, this is the biggest reinvention that you and I have seen,” Huang noted.
“This is incredibly exciting technology - you want to engage it. The impact to your company is incredible. And you want to be an early adopter.”
“This is the beginning of a decade of transformation. But you don't want to be second - this is the time, and you want to be first.”
You might also likeJony Ive, who famously designed the iPhone (among other iconic Apple devices), is about to become the design lead for OpenAI, the chatCPT AI giant that, for now, does not make a single hardware device.
The Wall Street Journal on Wednesday reported the impending deal, which sees OpenAI acquire Ive's io company in a deal valued at $6.5 billion. As part of that, Ive becomes the design lead for OpenAI, a role he's been slowly-stepping into for some time.
Ive, who famously led Apple's design for decades, left the company in 2019 and, in recent months, has expressed some misgivings about the possible negative impact of the previous products he's worked on (which might include the iPhone).
"I think when you’re innovating, of course, there will be unintended consequences, You hope that the majority will be pleasant surprises. Certain products that I’ve been very, very involved with, I think there were some unintended consequences that were far from pleasant,” said Ive earlier this month, according to the Verge.
While reports indicate that Ive and OpenAI CEO Sam Altman are interested in building AI-capable consumer hardware, a smartphone is probably not on that menu.
Instead, most expect the duo to focus on wearables like earbuds and smartwatches that could be enhanced with, for instance, cameras that could see your surroundings and use onboard AI to help you act on and react to them.
A soft approachIve's focus will also apparently be on upgrading OpenAI software's visual appeal. So expect an infusion of Ive-ness on ChatGPT on mobile and the desktop (where it has a particularly techy or dev-friendly look), as well as on Sora and Dall-E interfaces.
In the latter part of his career at Apple, Ive was most responsible for stripping away skeuomorphism – making digital icons look like their real-world counterparts – across Apple's platforms. OpenAI's software doesn't suffer from the skeuomorphic scourge, but some could argue its overall look is less than elegant.
If you're curious if Ive's design skills are still up to snuff, just take a look at the updated Airbnb, which Ive's Loveform firm redesigned. Loveform, by the way, is set to remain a stand-alone company and will, according to The Wall Street Journal, work with OpenAI as a client.
The news must sting Apple a little bit. The company, which partnered with OpenAI to include ChatGPT access in Apple Intelligence, has not only failed to deliver its own generative AI, but is falling behind the industry in delivering a true, combined hardware/software AI experience.
Open AI CEO Sam Altman (Image credit: Getty Images / Tomohiro Ohsumi / Stringer)Hints of hardware to comeIt'll be fascinating to see what Altman and Ive cook up, and we already have some hints.
Altman announced the deal by tweeting that he's "excited to try to create a new generation of AI-powered computers." Taken literally, we might expect an AI PC from the team, but I think here Altman means "computers writ large" in that most intelligent consumer electronics could be considered computing devices.
The tweet was accompanied by a video featuring a conversation between Ive and Altman, in which Altman described developing "a family of devices that would let people use AI to create all sorts of different things."
Without disclosing the product, Ive revealed that "the first one we've been working on has almost completely captured our imagination." Further, Altman added that Ive handed him the device to take home. "I've been able to live with it and I think it's the coolest piece of technology that the world will have ever seen."
No matter what they're building, it's worth remembering that the road to AI hardware success is already littered with the rotting carcasses of failed ventures like Human AI. Regular people have not shown great interest in wearing AI hardware that doesn't align with their current fashion choices.
thrilled to be partnering with jony, imo the greatest designer in the world.excited to try to create a new generation of AI-powered computers. pic.twitter.com/IPZBNrz1jQMay 21, 2025
That said, there may be an opportunity for OpenAI, Ive, and Altman in the smart glasses space. It's the one AI-connected device area that appears to be showing some real signs of life. That's mostly down to Meta's efforts with Ray Ban Meta Smart Glasses, but also evidenced by the upcoming influx of Android XR competitors from Google partners Samsung, Warby Parker, and others. Some were announced this week at Google I/O 2025, and all of them will feature Gemini at their core.
OpenAI and ChatGPT may be leading in the generative AI space, but Google Gemini is close behind. And if Android XR partners can deliver stylish Gemini Smart Glasses this year, it could quickly vault Gemini into the lead. At the very least, this puts pressure on OpenAI to deliver something.
Is Jony Ive the secret sauce that will make ChatGPT AI glasses, earbuds, smart watches, and other consumer hardware possible and desirable? Maybe. OpenAI says we'll see their work next year. Just don't expect a ChatGPT Phone.
You might also likeThe Bear season 4 has an intense trailer ahead of its Hulu (US) and Disney+ (international) release on June 25.
Maybe I'm reading too much into things, but deciding to include the ever-stressful clock could indicate we're counting down to the finale.
There are also moments of Carmy (Jeremy Allen White) trying to remember why he loves what he does and approaching a meeting with his mother (Jamie Lee Curtis), so even if it isn't the final season, there's still plenty of drama ahead, and it seems to be building to something big.
Previously, I noted that The Bear made Emmys history for most comedy wins, but it's still the most stressful Hulu show on TV, and I'm sure season 4 will be no different.
Check out the new trailer below.
What is The Bear season 4 about?(Image credit: Hulu)An official plot for The Bear season 4 reveals: "Carmen 'Carmy' Berzatto (Jeremy Allen White), Sydney Adamu (Ayo Edebiri), and Richard 'Richie' Jerimovich (Ebon Moss-Bachrach) push forward, determined not only to survive, but also to take The Bear to the next level. With new challenges around every corner, the team must adapt, adjust and overcome."
There's still quite a lot to speculate on regarding one of the best Hulu shows, so we'll have to wait patiently to find out more about season 4. It won't be long, though, and much like the clock in the trailer, I'm counting down the hours.
Until then, why not check out the biggest Hulu movies to stream in May 2025 or 5 new Hulu movies with over 91% on Rotten Tomatoes to tide you over.
You might also likeUsers can now change compromised passwords directly in their Chrome browser, in just a few clicks. This is the promise given in a new Google blog discussing the extensive changes the company is bringing to user authentication and identity verification.
Most browsers already come with a (rudimentary form of) password manager, allowing users to generate strong passwords, store their credentials, and auto-fill them for speed and convenience.
Now, Google’s Chrome devs, Ashima Arora, Chirag Desai, and Eiji Kitamura, said the company is building on that foundation to “fix compromised passwords in one click”.
TechRadar Pro readers can get 60% off Premium Plans at RoboForm now!
New users can take advantage of RoboForm’s exclusive deal and get 60% off the Premium Plan. With this deal, you can get unlimited password storage, one-click login & autofill, password sharing, two-factor authentication for added protection, cloud backup, and emergency access for trusted contacts. To claim this deal, visit this link and sign up for the Premium Plan to lock in this huge discount.
Preferred partner (What does this mean?)View Deal
Changing passwords“Automated password change makes it easier for users to respond when their credentials are at risk,” the blog reads. “When Chrome detects a compromised password during sign in, Google Password Manager prompts the user with an option to fix it automatically. On supported websites, Chrome can generate a strong replacement and update the password for the user automatically. This reduces friction and helps users to keep their account secure, without hunting through account settings or abandoning the process partway.”
Passwords are still, by far, the most common and popular form of authentication. They are also the least secure form, as people tend to create weak, easy-to-guess passwords, tend to share them with friends, family and coworkers, or store them in insecure locations that hackers can easily access.
The community has rallied behind alternatives such as passkeys, biometric authentication, or physical security keys. Google is also working on all of these (and then some), but stressed that passwords were “still the world’s most common authentication method,” suggesting that it’s not abandoning the practice any time soon.
The full blog is a rather interesting read, discussing a unified sign-in experience, improved identity verification, and enhanced session security. You can read it in more detail on this link.
You might also likeNEO Semiconductor’s push into 3D X-DRAM memory marks an ambitious attempt to rethink DRAM design for the AI and high-performance computing era.
While the promises - stacked layers, enhanced bandwidth, and reduced power consumption - are impressive, the practicality and consumer accessibility of these technologies remain open to scrutiny.
With the company projecting that its most advanced modules could reach densities of up to 512GB, it’s hard not to ask: who is this memory really for?
Complex architectures with limited consumer impactAt the core of NEO’s approach is a vertically stacked architecture that mimics the structure of 3D NAND.
In NEO’s own words, the array is “segmented into multiple sectors by vertical slits,” with “word line layers connected through staircase structures.”
The company compares its 3D X-DRAM density to the current 0a-node planar DRAM’s 48GB and claims to reach 512GB, but the implication that such capacities will trickle down into mainstream consumer products seems tenuous at best.
The proof-of-concept chips are still in the early stages. NEO is currently developing a test version of the simpler 1T0C architecture, with the more complex, and more promising, 1T1C version planned for 2026.
The 1T1C variant utilizes IGZO transistors paired with a cylindrical high-k dielectric capacitor. It promises improved retention time, reportedly beyond 450 seconds, and supports stacking up to 128 layers.
With further refinements, including the addition of 5nm-thick spacers to reduce parasitic capacitance, NEO claims stacking could exceed 512 layers.
The 3T0C design, which incorporates dual IGZO layers, is geared toward in-memory computing and AI applications.
Still, NEO’s statements about eliminating the need for TSV and enabling up to 32K-bit bus widths raise eyebrows.
Such bandwidth sounds transformative, especially compared to the projected 2K-bit bus width of HBM4, but scaling this level of performance in real-world systems is a non-trivial task.
From a broader perspective, the DRAM market hasn’t shifted significantly in terms of cost-per-GB over the past decade. Despite some fluctuations, the downward trend slowed considerably after 2012.
One might expect the MacBook Pro, for instance, to ship with far more RAM by default today than it did a decade ago, but that hasn’t happened.
Even with some price drops - DDR3 vs. DDR5 comparisons show modest improvement - the advances haven’t been revolutionary.
Commodity pricing may fluctuate, but the overall curve has flattened. Forecasts suggest we may be near a low point before another upswing.
So while 3D X-DRAM may indeed deliver bigger, faster memory by 2026, it’s unlikely these 512GB modules will be available to consumers anytime soon.
More likely, such capacity and speed will be reserved for AI servers and enterprise systems, rather than everyday desktops or laptops.
You might also like