A single missile can cost millions of dollars and only hit a single critical target. A low-equity, AI-powered cyberattack costs next to nothing and can disrupt entire economies, degrading national power and eroding strategic advantage.
The rules have changed: the future of warfare is a series of asynchronous, covert cyber operations carried out below the threshold of kinetic conflict. Battles will still be fought over land, sea, and sky, but what happens in the cyber domain could have a greater bearing on their outcome than how troops maneuver on the battlefield.
We were always heading in this direction, but AI has proven a dangerous accelerant. The entire military industrial base must become fortified against these risks, and that starts with continuous, autonomous validation of its cyber security defenses.
The Case for Autonomous ResilienceToday’s adversaries, whether state-sponsored actors or independent cybercrime syndicates, are deploying AI-driven agents to handicap critical systems across the entire military supply chain. These attackers aren’t focused on headline-making digital bombs, but a slow attrition, applying continuous pressure to degrade functionality over time. They’re also working anonymously: AI-enabled cyberattacks are executed by autonomous agents or proxies, making attribution slow or impossible.
Consider a hypothetical attack on the U.S. Navy. The Navy depends on a vast, decentralized web of small and mid-sized suppliers for everything from propulsion components to shipboard software systems. While these systems and suppliers may coalesce into the most technologically advanced Navy in the world, their interdependence is almost akin to human biology, in the way that a hit to one system can thoroughly destabilize another.
An adversary doesn’t need to breach the Navy directly. Instead, they can launch persistent cyberattacks on the long tail of maritime subcontractors, degrading national capability over time instead of in one massive, headline-making blow.
Third-party vendors, which often lack the financial resources to properly patch vulnerabilities, may be riddled with unsewn wounds that attackers can use as an entry point. But major security vulnerabilities aren’t the only way in. AI-driven agents can autonomously compromise outdated email systems, misconfigured cloud services, or exposed remote desktops across hundreds of these suppliers.
The impacts of these attacks can look like “normal” disruptions, the result of human error or some missing piece of code: delayed component deliveries, corrupted design files, and general operational uncertainty. However, the ill effects accumulate over time, delaying shipbuilding schedules and weakening overall fleet readiness.
Emerging threatsThat’s not even accounting for sanctions. If equipment is damaged, and replacement parts or skilled maintenance teams are restricted, one attack has just crippled a nation’s chip manufacturing capacity—potentially for months or years.
These attacks also get smarter over time. AI agents are designed for continuous improvement, and as they sink deeper into a system, they become more adept at uncovering and exploiting weaknesses. The cascading damage limits recovery efforts, further delaying defense production timelines and dragging entire economies backwards.
Despite these emerging threats, most defense and industrial organizations still rely on traditional concepts of deterrents, built around visible threats and proportional response: think static defenses, annual audits, and reactive incident response. Meanwhile, adversaries are running autonomous campaigns that learn, adapt, and evolve faster than human defenders can respond. You cannot deter what you cannot detect, and you cannot retaliate against what you cannot attribute.
Facing such dire stakes, defense contractors must exploit their own environments before attackers do. That means deploying AI-powered agents across critical infrastructure—breaking in, chaining weaknesses, and fixing them—to achieve true resilience. If the window for exploitation narrows, and the cost of action rises. “Low equity” means little against a high chance of failure.
Leveraging AI in Proactive DefenseFighting fire with fire sounds simple enough, but there are serious risks involved. The same AI tools that bolster organizations’ defenses against smarter, more covert attacks can also create new vulnerabilities. Large language models (LLMs) may cache critical weaknesses in their model architecture, and third-party components that contribute to the models’ effectiveness can also introduce new vulnerabilities.
Any AI-powered security tools should undergo a comprehensive vetting process to identify potential risks and weaknesses. Model architecture and history, data pipeline hygiene, and infrastructural requirements–such as digital sovereignty compliance–are all factors to consider when augmenting security with AI-enabled tools.
Even the cleanest, most secure AI program is not a failsafe. Defenders that rely too heavily on AI will find themselves facing many of the same problems that plague their counterparts who use outdated scanners.
A mix of false confidence and alert fatigue from automated risk notifications can lead to missed critical vulnerabilities. In a national security scenario, that can lose a battle. That can lose a war. Real, attack-driven testing makes up for where AI lacks, and when used in tandem with it, creates an ironclad shield against AI-enabled adversaries.
Artificial intelligence is a boon for society and industry—but it is also a weapon, and a dangerous one at that. Fortunately, it’s one that we can wield for ourselves.
We list the best Large Language Models (LLMs) for coding.
This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
One of the best website builders, Framer, just introduced a new feature that lets users update websites directly on the live page. Called On-Page Editing, the new tool is designed to have anyone, not just designers, make changes to websites quickly and safely.
In a press release shared with TechRadar Pro, Framer said that with On-Page Editing, users can fix typos, update text, swap images, and even create new pages without opening the design canvas, or navigating the Content Management System (CMS). Perhaps more importantly - they can do it without relying on someone else to implement changes.
“This isn’t just about making edits easier,” said Koen Bok, CEO and co-founder of Framer. “It’s about unlocking a whole new way to collaborate. With On-Page Editing, we’re laying the foundation for websites where designers build the system, but anyone can contribute with confidence.”
Removing bottlenecksOn-Page Editing integrates directly with Framer’s platform, offering single-click editing that syncs changes to the project, instantly. Rich text formatting, links, lists, and CMS page creation can all be done visually, without switching interfaces, the company explained. Framer also said that since teams can submit edits for review, bottlenecks between marketing, content, and design departments could be removed altogether.
Framer says that its end-to-end control of the stack, from canvas, across CMS, to hosting, allows the kind of workflow that keeps design integrity intact, while still allowing non-technical staff to contribute, and in real-time, at that. It expects template creators to benefit from the new offering as well, since they’ll now be able to create more static designs, while the flexibility of the system will enable anyone to edit and publish without coding knowledge.
On-Page Editing is available immediately for all paid Framer plans. Prices range from $75/month/site, to $200/month/site. There is a also the option of custom pricing for enterprise clients.
You might also likeWith IT infrastructure growing more complex and teams under pressure to do more with less, it’s time for organizations to rethink their observability strategy before costs, burnout, and blind spots spiral out of control.
Across industries, legacy observability tools are buckling under the weight of today’s dynamic infrastructure. These traditional monitoring systems were designed for a world where environments barely moved, data trickled in manageable amounts, and collecting more metrics felt like progress.
But that era is long gone, and teams stuck in ‘collect everything’ mode are paying the price with runaway costs, spiraling complexity, and blind spots that turn small hiccups into full-blown outages.
In today’s fast-moving, containerized world, this strategy is backfiring. What once felt like a safety net has morphed into a data landfill, drowning teams in noise, burning them out, and surprising them with cost overruns that deliver the only visibility nobody wants: a meeting with the CFO to justify the bill.
The observability promise that fell apartFor years, engineering teams were sold a simple idea: more data meant more control. That advice made sense when infrastructure was static and applications evolved slowly – capturing everything often delivered the insights teams needed. But the rise of cloud technology changed everything, turning environments ephemeral and accelerating the pace of change and telemetry growth. Yet many teams still cling to the old ‘collect everything’ strategy, even as it drags them down.
Modern systems don’t wait. They scale instantly, shift constantly, and produce overwhelming volumes of telemetry. The tools that once brought stability are falling behind; they weren’t built for today’s level of scale or complexity. They’ve become rigid, noisy, and expensive, and the cracks are starting to show.
In sectors like aviation, even brief outages can result in millions of dollars in losses within minutes. Elsewhere, the fallout is just as real: frustrated customers, eroded trust, and reputational damage.
What once felt like a smart investment has quietly become a liability. Many organizations are waking up to the uncomfortable truth: their observability stack is no longer fit for purpose. Instead of becoming the true utility teams can rely on, it adds to the technical debt they’re actively trying to mitigate.
When teams can’t separate signal from noise, dashboards become cluttered with irrelevant metrics, alerts never cease, and real issues slip through the cracks. This constant stream of distractions imposes a steep distraction tax: every context switch, every false alarm, every hunt for meaning chips away at an engineer’s productive time and mental energy.
Over time, this chaos breeds reliance on tribal knowledge from a few seasoned ‘heroes’ who know where the bodies are buried. These heroes become the crutch that props up the system, celebrated for their late-night saves. However, a hero culture comes at a high price, with burnout, a lack of knowledge sharing, and stalled innovation, as teams spend more time firefighting than building differentiating features.
Observability should enable innovation, not kill it. When engineers are drowning in data without clarity, the best they can do is react. And in a world moving this fast, organizations that can’t move past constant triage will find themselves leapfrogged by the competition.
What does good observability look like?Solving this problem isn’t just about new tools – it demands a strategic approach to your business pain. A strong observability strategy helps you deliver a better customer experience, enhance employee productivity, and increase conversion rates and revenue.
It delivers clear insights into the performance of your digital investments by revealing feature adoption trends, capacity and scaling gaps, and release quality and velocity issues. Done right, observability fuels a culture shift where teams embrace it as an enabler, not another distraction tax.
A clear telemetry collection methodology is essential to make observability a strategic asset rather than an operational burden. This methodology should be guided by well-defined Service Level Objectives (SLOs) and error budgets, which set the standard for what matters most to your business and customers.
By aligning telemetry collection with these objectives, you ensure your observability strategy surfaces only the data that helps measure and improve outcomes. This disciplined approach connects engineering efforts directly to business value, enabling teams to confidently invest in features, optimize performance, and scale systems without getting lost in the data deluge.
Even the best telemetry strategy will fail if observability is treated as an afterthought or siloed concern. Successful organizations make observability a shared responsibility by embedding it into team norms, workflows, and incentives. That starts with clear executive sponsorship to set expectations, coupled with training that gives every engineer confidence in reading, interpreting, and acting on telemetry data.
Organizational Change Management (OCM) practices help teams adopt observability incrementally, shifting the culture from reactive heroes to proactive, data-driven improvement. When observability becomes part of how everyone builds and operates software, it transforms from a distraction into a force multiplier for innovation and resilience.
Observability done right isn’t optional – it’s a competitive advantage. Teams that treat it as a strategic utility, guided by clear objectives and supported by a culture of shared responsibility, will outpace those stuck in reactive firefighting.
Now is the time to rethink your observability strategy: invest in disciplined telemetry, align it with what matters most to your business, and empower your teams to build with confidence. Strong observability turns bold strategies into market leadership and keeps your teams focused on the future.
We've listed the best business intelligence platform.
This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
While unmanned aerial vehicles have been in operation for nearly 200 years, drones, as we know them today, were first used by the military in 2006 before technological innovations turned them into in-demand children’s toys.
Now, industries, including mining, oil and gas, ports, utilities, and public safety teams, are starting to recognize their capabilities. When used for surveillance or inspection, they can protect workers from entering hazardous areas and deliver greater situational awareness about operations or incidents when every second counts.
Today, the industrial drone sector is rapidly evolving, with 2025 marking significant advancements in technology and applications across various industries. Several countries are making drones even more attractive for industrial use, allowing them to be flown beyond visual line of sight (BVLOS) at low heights close to buildings and infrastructure.
Drones for safer, more efficient operationsThe global drone market is projected to grow from $36.7 billion in 2024 to $44.32 billion in 2025, with industrial applications being a significant contributor to this growth. This has the potential to unlock new Industry 4.0 use cases for operational efficiency and worker safety improvements.
For example, at vast mining sites, drones can handle site surveys more safely, effectively and sustainably than on foot or by plane. They can be dispatched multiple times, equipped with payloads including LiDAR sensors, HD cameras, magnetometers and thermal imaging cameras to gain greater intelligence or terrain mapping
Once a mine is operational, drones can inspect waste and stockpiles, relieving workers, often in the harshest of temperatures. They can also more accurately and rapidly calculate stock volumes for better management. Operational disruption and worker hazards are reduced when drones are flown to inspect the impact of blasting.
Using drones to handle perimeter surveillance at large industrial sites reduces reliance on people, vehicle wear and tear and fuel consumption which contributes to industrial enterprises sustainability efforts. Oil and gas companies can use them to inspect storage tanks, pipelines, cooling towers and substations.
Dispatching drones from one or multiple operation centers ahead of first responders enables early assessment of a situation's severity, helping to support efficient resource allocation while also receiving AI-powered intelligence that can inform and accelerate decisions to help keep workers, property, and premises safer.
For instance, in a wildfire, drones can be flown lower than a manned aircraft, using thermal cameras to identify heat spots through the smoke. Using data accessed through drone flights, teams are better equipped to handle these situations and protect themselves, the community, and the environment. Today hundreds of US police departments have drone programs. Skilled police pilots operate drones for search and rescue missions, crowd and event monitoring, and various other tasks.
Integrating edge and AI technology for reliable drone flightsConnected to an edge and AI platform, drones can benefit industrial enterprises with automated activities that enhance efficiency and safety.
For example, leveraging AI, analytics and machine learning, changes in sensor data could trigger a drone flight – whether to monitor a perimeter breach, an equipment malfunction, or to notify the correct first responder teams. Using data and video from those drones, any team will be better informed, ensuring that the right people get to where they need to be faster with the right equipment.
To do this, industries must deploy an integrated digitalization platform that uses the right mix of technologies. That means reliable connectivity, real-time edge data processing, ruggedized drone hardware and an extensive range of software to enable new use cases.
To fly beyond BVLOS, connectivity must be robust, meaning seamless handovers at speed. Using Wi-Fi alone, as networks become overloaded, handovers between access points can be delayed, meaning that drones fail. However, by implementing a platform that integrates multiple technologies, including public and private 4G and 5G wireless and Wi-Fi, enterprises will benefit from redundancy and reliable connectivity even as the drone flies BVLOS.
An on-premises industrial edge processing solution allows data to be processed in real time and consumed by applications for new efficiencies. Real-time data will better inform teams of the latest situation, while enhancing productivity and ensuring the right people and equipment are in the right place at the right time.
Drone hardware must be ruggedized, built to withstand harsh industrial environments and weather conditions at sites including mines, agricultural areas and oil and gas facilities. A drone solution that can be used with a mix of payloads and software will unlock flexibility for any enterprise. Efficiency is enhanced further using drones that are charged remotely and flown further BVLOS.
The drone use cases will vary depending on the industrial enterprise's needs, but in most cases, they will allow them to achieve multiple goals simultaneously. For example, industries can access more accurate data for operational improvements while also protecting their workers.
We list the best Enterprise Resource Planning (ERP) software.
This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
Super Apps are gaining traction globally, combining shopping, banking, messaging, and more into seamless mobile experiences. These platforms are reshaping how consumers live, engage, and transact, particularly in mobile-first environments.
At the same time, global economic shifts, including rising tariffs, geopolitical uncertainty, and growth plateaus in traditional markets, are prompting brands to explore opportunities in new markets. A pattern is emerging, with mobile-first businesses pushing to expand beyond their home turf, not just for growth but to reduce exposure to risk.
I refer to this emerging trend as internationalization. It is not a theoretical concept, but a real and active shift that we are already seeing play out. Brands are entering and testing new regions through mobile channels, especially in places where Super Apps have already become the central gateway to mobile life. Mobile apps offer a fast, scalable way to reach new audiences, particularly where mobile commerce dominates.
Understanding the intersection of these trends is crucial for anyone helping to shape how global brands will continue to grow.
Why internationalization is acceleratingInternationalization is no longer just a growth strategy. It is becoming a necessity. As trade restrictions increase and supply chains become more fragile, relying on a single market creates vulnerability. Regulatory shifts, currency swings, and platform volatility can all impact performance overnight.
Recent advertising patterns make this clear. In response to new U.S. tariffs, platforms like Temu and Shein have adjusted their strategies by pulling back ad spend in the region and ramping up investment elsewhere. Shein has reportedly increased its ad spend by more than a third in the UK and France, while Temu boosted its budgets by 40% in France and 20% in the UK.
Expanding into new regions spreads risk and opens access to mobile-first populations with advanced digital behaviors. And for good reason. Localizing an app experience is one of the fastest and most effective ways to test, learn, and scale.
The benefit is not just about protection. It is also about growth through diversification, reaching new audiences, unlocking cultural insights, and building relevance beyond your base market.
Super Apps as the platform advantageIn many high-growth regions such as Southeast Asia, Latin America, and South Asia, Super Apps are not just part of the consumer journey. They are the consumer journey. These platforms function as commerce, payment, and engagement ecosystems, often all within a single app.
Even in markets that are yet to fully embrace Super Apps, platforms are increasingly bundling services and moving toward more integrated, mobile-first experiences. For brands, this represents an opportunity to meet users where their digital lives are already converging and to shape that convergence in new regions.
Uber is a prime example of this evolution. It has moved far beyond ride-hailing to offer food delivery, groceries, freight, and even financial services − all from a single interface. This expansion has helped power Uber’s advertising business, which surpassed $1 billion in annual revenue in 2023, driven by rich first-party data and high-frequency user interactions.
For brands looking to expand internationally, these platforms offer direct access to engaged, mobile-native audiences. But success takes more than presence. It requires thoughtful localization aligned with local preferences, payment behavior, and platform expectations.
That might mean adapting checkout flows to local payment providers, surfacing regionally relevant promotions, or aligning creative with cultural shopping habits; all of which depend on having access to the right regional insights and data. When brands get this right, they can scale faster and more sustainably across markets that might otherwise be disrupted by economic or policy shifts.
Super Apps offer reach, but they are not plug-and-play. Brands need to treat them as local infrastructure, shaped by how people shop, pay, and engage in each market. Success depends on designing for that context, not just adapting language or layout.
Measurement with trust and depthAs Super Apps bring together commerce, content, communication, and payments, they create both a challenge and an opportunity for measurement. The customer journey becomes more consolidated but also more complex. Traditional attribution models often fall short when interactions span multiple functions within one app environment.
At the same time, these platforms offer something rare. Because users stay logged in and conduct a wide range of activities within a single ecosystem, Super Apps create a unified stream of high-quality, first-party data. When managed responsibly, this data can offer brands a richer, more accurate understanding of what drives performance across the entire customer lifecycle.
But this opportunity requires a shift in how measurement is approached. Brands need to move beyond fragmented analytics and build a framework that accounts for the full breadth of user behavior. This is not just about reporting. It is about harmonizing data across channels and touchpoints to create a more complete view of impact.
Equally important is how that data is handled. Privacy must be built into the infrastructure, not layered on after the fact. When a user's entire digital experience takes place within a single app, privacy is not just a compliance issue. It becomes an essential commitment.
For marketers operating inside Super Apps, measurement is no longer a technical task. It is a strategic imperative. The brands that win in this environment will be those that can gain from insights without compromising integrity.
Navigating complexity with strategic intentSuper Apps deliver scale but also come with complexity. Closed ecosystems can restrict access to data and reduce flexibility. Platform dependency raises questions about control and long-term resilience.
This is not a reason to hold back. It is a reason to engage strategically, with visibility, adaptability, and a clear approach to measurement and integration. Sustainable growth means knowing how to work within these environments while maintaining your own strategic independence.
Global expansion today is not about replicating what worked at home. It is about adapting, aligning, and operating with a deep understanding of where you want to grow next.
Compete with contextSuper Apps are not just a trend. They are becoming the infrastructure for mobile-first economies and the entry point to digital life in many parts of the world. At the same time, internationalization is accelerating as brands seek new revenue, new users, and greater resilience.
Those who succeed will not just be available in a region. They will localize with purpose, measure with precision, and deliver value within the platform experience itself.
We've listed the best mobile payment app.
This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
The potential benefits and opportunities from AI agents in both our business and personal lives center around one word: productivity. Being able to do more with the help of AI seems like a no-brainer and a road to a less busy and hectic life.
But we’ve heard this before, haven’t we?
Like the introduction of the internet, smartphones, and app stores – all had the promise to simplify our lives and boost productivity. While they have had a profound impact on the way we work, it hasn’t changed how busy we are. In fact, Pew Research found that 47 percent of office-based employees said technology has actually increased their work hours. We always seem to fall back into the “busyness trap,” using technology to get things done faster, but filling our newly-found spare time by adding more menial tasks to our lists.
Why? Because we live and work in a world where being busy is equated with success. However, this new AI era presents a chance to approach things in a completely different way.
The fascination around AI being able to unlock new levels of productivity is absolutely warranted – it is an opportunity to help employees save time on administrative tasks and focus on delivering more strategic value through bigger, more impactful projects.
But in order to achieve that, we need to spark a mindset shift around what productivity truly means and to become comfortable with a “fewer but better” philosophy. Otherwise, we’re setting ourselves up to repeat the same busyness cycles we’ve seen before.
Business leaders now have the chance to rethink their approach, create an employee experience built on impact vs. to-do lists, redefine success standards, and optimize an AI-driven workforce. Here are a few ways to get started:
It’s time to redefine productivityBy automating routine and tedious tasks, AI adoption is opening the door to new levels of productivity, but not in the sense of checking more boxes. Real productivity means doing better, bigger, and more valuable things with less.
This requires a cultural shift. Leaders must be willing to rethink what productivity actually is. Allowing people to focus more time on more meaningful initiatives can result in higher value and better execution across their organizations. By fostering a culture that emphasizes meaningful results, quality outcomes, and impactful contributions, leaders can inspire employees to align their efforts with the broader goals of the organization.
Break the mold: normalize not being busy“There are not enough hours in the day.” “I am working as hard as I can.” “I wish I could, but I’m just too busy.” We all know these familiar phrases because more often than not, we are all busy being busy.
Busy people are too often seen as important people, but with the help of AI, we don’t have to be so busy to be accomplished. AI offers an opportunity to focus on essential, strategic, and high-value work, but it’s all for naught unless we can alter the perception around not being busy. We have to move away from the idea that busyness is a badge of honor and instead embrace intentionality, balance, and strategic prioritization.
Look for ways to reduce the noise and sit in silenceOne of the key opportunities of integrating AI into workflows is letting AI agents handle the “noise.” All of those small yet time-consuming tasks that are keeping us busy can be automated. Some examples include data entry, scheduling and calendar management, time zone coordination, invoice generation, and more. Letting AI eliminate the noise of our everyday seems easy enough, as long as we don’t replace them with more tedious tasks.
Being OK with allowing AI to take over the mundane and not replacing that time with other filler tasks is a skill business leaders will need to adopt, foster, and encourage their teams to exhibit as well. It requires continuously re-evaluating the essential vs. non-essential tasks, and assigning the non-essential to AI agents. A day with three strategic priorities completed can be just as – if not more – successful as a day with ten or more to-dos checked off.
Move away from multi-taskingWe’re now so used to constant context switching between devices, screens, conversations, and projects. The idea of being productive has centered around accomplishing multiple things, and that is traditionally done by multi-tasking. Although constantly switching between tasks may seem productive, it often results in mental exhaustion, diminished concentration, and a lower quality of work.
But with the growing adoption of AI agents, we have the opportunity to lean into hyper-focused productivity and hone in on specific initiatives that require and deserve our undivided attention. Over time, this focused approach is what drives more meaningful outcomes, greater personal fulfillment, and material improvement in productivity and effectiveness.
Embrace a better way forwardDespite the potential of past technological advancements, the reality is they haven’t alleviated our busy lives as we hoped. However, AI agents offer an incredible opportunity, not just to do more, but to do better. But business leaders need to be willing to redefine cultural and workplace norms.
By focusing on quality over quantity, reducing the constant noise, and moving away from multi-tasking, we can unlock the real promise of AI and create a future where productivity energizes rather than exhausts the workforce, all resulting in better business outcomes.
We've listed the best productivity apps for iPad.
This article was produced as part of TechRadarPro's Expert Insights channel where we feature the best and brightest minds in the technology industry today. The views expressed here are those of the author and are not necessarily those of TechRadarPro or Future plc. If you are interested in contributing find out more here: https://www.techradar.com/news/submit-your-story-to-techradar-pro
Fubo is the exclusive home network of the EPL in Canada. Every single game of the 2025/26 Premier League season is available to stream on Fubo and you can watch them all online from anywhere with a VPN.
Season dates: August 15 2025 – May 24 2026
Stream every match: Fubo
Download a VPN to watch overseas
Hoping to defend their title for the first time in more than 40 years, Liverpool have splashed the cash in the off-season, with Hugo Ekitike, Florian Wirtz, Milos Kerkez and Jeremie Frimpong all arriving at Anfield. The Reds look exceptionally strong but will face a stern challenge from the likes of Man City, Chelsea and Arsenal.
It’s also a big season for Man Utd who have signed a new front three. If they start poorly, then manager Ruben Amorim could be in trouble. Tottenham will also be hoping to build on their Europa League success and will expect a much-improved campaign under new boss Thomas Frank.
Newcastle and Crystal Palace both won trophies last season and will hope for more success in 2026, while it will be fascinating to see if Leeds, Burnley or Sunderland can avoid the drop straight back down to the Championship.
You won’t want to miss any of the action, so here’s how to watch every Premier League live streamson Fubo in Canada.
How to watch Premier League on Fubo in CanadaIt's nice and simple for Canadian soccer fans with Fubo streaming all 380 Premier League games on their platform!
However, you will need one of the sports plan to gain full access of Fubo's Premier League coverage. These packages are as follows:
Sports Monthly: CA$31/month (CA$4.50 off your first month)
Sports Quarterly: CA$27.99/month (CA$15.75 off your first quarter)
Sports Annual: CA24.50/month (CA19.25 for your first year)
Stuck abroad and looking to access your Premier League streams? Use NordVPN to trick your streaming device into thinking it's back in the Great White North.
How to watch Premier League on Fubo from anywhereFubo is only available to watch in Canada, which means it won't work if you try to stream EPL soccer when overseas on business or vacation. However, you can always download a VPN to overcome these regional restrictions and access Fubao as normal when you're outside Canada. You may be surprised how simple it is to do.
Watch Premier League on Peacock from anywhere with one of the best VPNs:
Editors ChoiceNordVPN – get the world's best VPN
We regularly review all the biggest and best VPN providers and NordVPN is our #1 choice. It unblocked every streaming service in testing and it's very straightforward to use. Speed, security and 24/7 support available if you need – it's got it all.
The best value plan is the two-year deal which sets the price from $2.91 per month including four months free. There's also an all-important 30-day no-quibble refund.
- Try NordVPN 100% risk-free for 30 daysVIEW DEAL ON
Fubo Q+AWhat streaming devices are supported by Fubo?Web Browsers (Chrome, Firefox 115, Edge, Safari)
Mobiles and tablets (Android 7.0 or above, iOS 15 or above)
Amazon Fire TV
Android TV
Apple TV
Chromecast (2nd Generation or above)
Google TV
Hisense VIDAA (2021 devices and newer)
LG Smart TV (LG WebOS 3.5 or above)
PlayStation (PS4, PS5)
Roku (2, 3 & 4, Streaming Stick, Express/Express+, Premiere/Premiere+, Ultra/Ultra LT)
Samsung Smart TV (2017 models and later
VIZIO (SmartCast TV 2016 and newer)
Xbox (One, X, S)
Xumo
In addition to Premier League soccer, a subscription to Fubo gets you access to a wide range of sports content.
Soccer fans will be able to watch other leagues, such as Serie A from Italy and Ligue 1 in France. Fubo also carries Canadian men's and women's national team matches, along with the Canadian Premier League.
Fubo also features a broad selection of other sports content, including NBA TV and access to more than 10,000 live sporting events annually. Subscribers can also enjoy over 25,000 TV episodes and movies as video-on-demand each month.
We test and review VPN services in the context of legal recreational uses. For example:1. Accessing a service from another country (subject to the terms and conditions of that service).2. Protecting your online security and strengthening your online privacy when abroad.We do not support or condone the illegal or malicious use of VPN services. Consuming pirated content that is paid-for is neither endorsed nor approved by Future Publishing.
Google Gemini will remember all those questions you ask on the app now, unless you ask it politely not to. The AI assistant will look at and reference your past chats as a way to personalize future ones, though only for Gemini 2.5 Pro for now.
With the memory function, Gemini could theoretically recall your favorite party themes, the YouTube channels you're into, or how you like to compose emails to your friends, without you having to remind it. Gemini's memory is not dissimilar to the memory feature offered by ChatGPT and other AI chatbots. And while ChatGPT can now connect to your Google account, Gemini has native access to Gmail, Calendar, and Google Docs.
The memory is on by default, so you'll have to make a slight effort if you'd rather Gemini not track your chats. You can turn the memory off or back on again in the settings menu under personal context, sliding the “Your past chats” option to off.
In addition, the “Gemini Apps Activity” section is now called “Keep Activity.” When that setting is active, some of your future uploaded files and photos may be used to train and improve Gemini and other Google services. You can opt out of that, too, if you wish.
Temp talks(Image credit: Google)As much as Google is eager to make sure Gemini can remember what you say to it, the company paired the rollout with its opposite, the new Temporary Chat. This incognito mode makes every conversation a one-off, and makes sure it isn't saved to your history, appears in the activity list, or sent to Google to train Gemini.
After a 72-hour hold for safety, they're they’re deleted completely, unless you submit explicit feedback, which will be processed and then discarded. Temporary Chat is designed for those moments when you want to ask something you’re not sure you’ll want remembered, whatever that might be.
The temporary chats might reassure people worried that Google just wants more of their data, but the company clearly hopes people find value in giving Gemini a memory. Google's goal is to encourage people to think of the AI as a long-term conversational partner, not just a tool you reintroduce yourself to every time you interact with it.
Regardless of any concerns, AI developers are very keen on memory, since, without it, AI assistants are just clever parrots. Technically, they still are just parroting data even with personalization, but they'll be less clever about it. Whether that’s exciting or unsettling depends on your appetite for intimacy with software.
You might also likeThinking of building a new computer? What if I told you that you could get the PC case for free - you probably wouldn't believe me, would you? Well, you can - with a couple of notable caveats, but this is still fundamentally a freebie.
The Verge reports that Teenage Engineering (more typically associated with nifty audio hardware) now has a sequel to its Computer-1 PC case. Can you guess what it's called? Yep: the Computer-2 case - in a stunning stroke of creativity.
The key difference is that, unlike the original case, which cost $ 149 in the US, the follow-up is free. Another notable change is that instead of aluminum, it's made of plastic, which is obviously a cost-saving measure.
In fact, it's a single sheet of semi-transparent plastic that folds into a PC chassis (using hinges and snap hooks). There are no screws needed, and the motherboard can simply be pushed and clicked into place, Teenage Engineering informs us.
It must be a (small) mini-ITX motherboard, mind, as this is a small form-factor (SFF) case for a compact PC build. It can fit an SFX power supply and a dual-slot graphics card (7 inches is the maximum length for the discrete GPU).
As well as the sheet of plastic that forms the chassis itself, you get the bits and pieces to go with it in terms of various power cables, fasteners, feet for the case to stand on, and so forth.
Analysis: What about those catches?(Image credit: Teenage Engineering)Okay, so coming back to the cost being a – literally – round figure of nothing - how on earth can this PC case be provided for free? Well, it can't, and Teenage Engineering is charging for the shipping to cover sending out the Computer-2 as you might guess (and likely a bit of that is to cover the cost of production). Of course, that's fair enough - it does mean you are getting the components themselves for free.
The other catch is that Computer-2 is already marked as sold out on the official website; you can only click to receive a notification when stock is back in. You can still examine the instructions, though, to get a better understanding of what construction involves, and exactly what you'll get.
By all accounts, the Computer-1 is a great case for a small PC build - and this new plastic take on the formula looks equally cool, especially given the price here (or lack of it). Anecdotally, I've seen a report of the shipping costing $16 (in the US) if you're wondering exactly what you might end up paying (when it's back in stock).
As for mulling over buying a cheap case in general, you can get some great products at the budget end of the market. Just be careful around airflow issues - especially if it's a gaming PC that you're building - and be sure to consult reviews to ensure you're not picking up a dud (and also our best PC cases buying guide will come in handy).
You might also likeOpenAI has started to roll out Gmail, Google Calendar, and Google Contacts integration into ChatGPT for Pro and Plus users today.
As a ChatGPT Plus subscriber, I was really looking forward to being able to connect my Gmail inbox to ChatGPT, as OpenAI CEO Sam Altman and the team demonstrated at the recent launch of ChatGPT-5.
Imagine my disappointment then, when, after going into ChatGPT’s Settings and connecting it all up to Google (a process which involves going through several permission screens), I found out that OpenAI is reserving the Gmail search in chat for ChatGPT Pro users.
Pro vs PlusA ChatGPT Pro subscription costs $200 (£200 / AU$307 ) a month, compared to my $20 (£20 / AU$31) Plus subscription, just to use a feature that I get for free with my Google Gemini account!
Since Gemini and Gmail are both made by Google, I can understand that it must be a lot easier for Google to connect them, but I can’t quite understand the logic of putting the feature behind a $200 paywall for ChatGPT users.
I actually use Gemini to search my Gmail all the time - it’s much, much better than the standard search feature in Gmail because you can type in somewhat vague queries like “find me the email where I talk about trains and stamp collecting” and it does a great job of finding the exact email, even if you’ve no idea who it was from or when you sent it.
Having that feature in ChatGPT would have been great, but it seems it’s not going to happen for me anytime soon.
Deep Researching my GmailThere is one caveat: On a ChatGPT Plus account, you can search Gmail through the Deep Research tool. Deep Research is OpenAI’s powerful research tool that’s meant to be used to research complex questions that require a full report to answer, not for simple things like searching your Gmail for a missing email.
For purely searching through your Gmail, I would say it works, but it is functionally useless: You click on Deep Research, then make sure that Gmail is turned on as one of the sources, then type in your query, and Deep Research ponderously goes off on its research process, which can take minutes.
It eventually comes back with good results, but nobody has got 5-10 minutes to wait for a simple Gmail search. I’d rather just go into Gmail and use the standard search bar and save myself minutes of time, or I’d use Gemini, which still seems like the best option for now.
You might also likeThe whole idea of an AI phone is a rather new one, but I’d argue that Google has been both the driver and frontrunner of phones that put artificial intelligence at their center.
This started with the Google Pixel 6 and its Tensor chip, which was built around delivering strong neural network machine-learning performance rather than the best CPU and graphics processing speeds. Then, with the Pixel 8, we saw Google really embrace AI features, especially those with generative capabilities like the Magic Editor.
While some might argue that the Samsung Galaxy S24 was the first ‘AI phone’, I’d say the Pixel 8 was the one that got the ball rolling, especially as Samsung used Google tech to underpin some Galaxy AI features.
This was all built upon with the Google Pixel 9 and its stablemates, which all felt like they put AI at the heart of the Pixel experience, rather than just offer tools on top of a standard smartphone user interface.
But with the Samsung Galaxy S25 and Galaxy Z Fold 7 landing with improved AI features, and Apple Intelligence starting to get up to speed and offer a solid suite of features on compatible iPhones, as well as some synchronicity with Macs, Google’s AI phone crown could be up for grabs.
So I feel that for Google to keep ahead and continue to blaze a trail for practical, and hopefully safe, consumer use, it'll need to bring more to the table with the Pixel 10.
Accelerated AI(Image credit: Future / Google)But what would that look like? Well, that's a good question.
In some ways, Google already offers up a good suite of AI-powered tools, from smart call handling and document summaries to real-time translation and photo editing tools that can rework entire shots. And a lot of these are easy to access and use, something which helped make the Google Pixel 9 Pro our phone of 2024.
However, there’s room for more. Ideally, I’d like to see some of these smart AI tools work across the whole of the Pixel Launcher experience.
For example, I’d like to use Gemini to summarize what I’ve got on my Pixel phone, say, telling me apps that I’ve not used for ages or games that I’ve not played and may have forgotten about.
And I’d be very keen to have the summarization tools built out to provide me with snippets of information in third-party apps, say to quickly summarize a chapter of an ebook so I can easily pick up where I’ve left off without reading back over stuff for a refresher.
More than anything else, I’d love to have audio readouts of all kinds of text in a voice as close to natural language speaking as possible. I listen to a lot of podcasts, especially when I'm commuting or cooking, but this does mean I can fall behind on a backlog of reading.
To tackle this, I’d love to be able to ask a system-level AI to read aloud a magazine feature either from a digital source or scanned in via a phone’s camera. I’ve stumbled across some AI tools that can sort of do this, but none deliver what I’m looking for and in a seamless way that AI luminaries tend to tout.
This isn’t a pipedream, as Google already has NotebookLM with its Audio Overviews that can make a podcast out of various inputs. So if this tech could be baked into the next-generation Pixel Launcher or the rumored Pixel 10 phones, it would be a boon for me and likely other people looking for aural satisfaction.
Ultimately, while I appreciate the creative opportunities of some AI tools in terms of generating images or deeply manipulating photos, I’d like AI in general to be more about making my daily life easier and helping me better digest the almost overwhelming amount of content there is to consume in the virtual and real world.
I have faith that Google is one of the companies best placed to do this, even with my nervousness about how much the search giant’s algorithms already influence the dissemination of information.
But I do think it needs to use each new generation of Pixel phone to keep pushing the envelope, while the likes of Apple and Samsung appeal to the everyday phone users, not just to keep the AI phone crown, but also to push AI innovation forward in a way that’s genuinely useful and productive for people.
You might also likeA mysterious death; a young hotshot lawyer looking to make his mark; tense courtroom battles –The Rainmaker TV series certainly has all the hallmarks of a classic John Grisham thriller. US viewers can tune into The Rainmaker from Aug 15 but read on and we'll show you how to watch The Rainmaker online from anywhere with a VPN.
Release date: Friday, August 15
TV channel: USA Network
US stream: Sling TV / Fubo / Peacock (week later)
Use NordVPN to watch any stream
Rudy Baylor (Milo Callaghan) is the drama's main protagonist, just out of law school and looking for his first big case. He soon finds it when he picks at scabs he shouldn't, delving deep into the suspicious circumstances around an unexplained fatality and coming face-to-face with the formidable courtroom opponent of Leo F. Drummond (John Slattery).
Previously adapted for the silver screen by Francis Ford Coppola and starring Matt Damon, USA's TV version of the iconic 1997 legal drama will take its time retelling this complex tale of conspiracy and corporate wrongdoings over 10 tense episodes.
Grab your gavel and prepare to watch The Rainmaker online and stream weekly episodes from anywhere.
How to watch The Rainmaker online in the USYou can watch The Rainmaker on USA Network on Fridays at 10pm ET/PT in the US. New episodes air weekly starting with the premiere on Friday, August 15.
Don’t have cable? You can also watch USA via Sling TV on its Blue plan costing from just $45.99 per month or $4.99 per day. Or, with more than 200 channels, Fubo is a more comprehensive cord cutting alternative. It costs from $84.99 a month and new users can get a free Fubo trial.
The Rainmaker will also be available to stream on Peacock, with episodes landing on the platform one week after they go out on USA. The Peacock price starts at $10.99 a month and you can get 12 months for the price of 10 by opting for an annual plan.
Away from the US? Use a VPN to watch The Rainmaker on Sling, Fubo or Peacock from abroad.
How to watch The Rainmaker online from outside your countryIf you’re traveling abroad when The Rainmaker airs, you’ll be unable to watch the show like you normally would due to annoying regional restrictions. Luckily, there’s an easy solution.
Downloading a VPN will allow you to stream online, no matter where you are. It's a simple bit of software that changes your IP address, meaning that you can access on-demand content or live TV just as if you were at home.
Watch The Rainmaker from anywhere with one of the best VPNs:
Editors ChoiceNordVPN – get the world's best VPN
We regularly review all the biggest and best VPN providers and NordVPN is our #1 choice. It unblocked every streaming service in testing and it's very straightforward to use. Speed, security and 24/7 support available if you need – it's got it all.
The best value plan is the two-year deal which sets the price from $2.91 per month including four months free. There's also an all-important 30-day no-quibble refund.
- Try NordVPN 100% risk-free for 30 daysVIEW DEAL ON
How to watch The Rainmaker online in CanadaJust like south of the border, Canadian viewers can watch The Rainmaker on the USA Network. The difference is that episodes will go out a day later on Saturdays at 10pm ET/PT.
Rather stream the show? Use the CTV.ca website. You'll likely need to enter your cable provider details.
US viewer currently traveling in Canada? Download a VPN to connect to your streaming service back home and watch The Rainmaker no matter where you are.
Can I watch The Rainmaker online in the UK?At the time of writing, there has been no confirmation that The Rainmaker will be shown in the UK.
Currently visiting the UK? Downloading a good VPN will let you effortlessly connect to your usual streaming services from anywhere in the world.
How to watch The Rainmaker online in AustraliaTo stream The Rainmaker in Australia, you'll need a Stan subscription. Episodes will land on Saturdays, with the first one available on August 16.
Stan subscriptions costs from $12 a month for a Basic plan, with incremental increases if you want to add HD/4K quality and additional devices.
Or if you’re visiting Australia from abroad and want to watch on your home service, simply download a VPN to stream The Rainmaker just as you would back home.
Can I watch The Rainmaker for free?The show isn't on any free services, but US viewers can use the Fubo 5-day free trial to watch available episodes of The Rainmaker for free.
The Rainmaker trailerThe Rainmaker castThere's set to be 10 episodes of The Rainmaker.
(Dates all US premieres)
VPN services are evaluated and tested by us in view of legal recreational use. For example:a) Access to services from other countries, (subject to the terms and conditions of that service).b) Safeguarding your online security and making your online privacy more robust when abroad.Future plc does not support nor condone the illegal or malicious use of VPN services. We do not endorse nor approve of consuming pirated content that is paid-for.
Researchers from the American Psychological Association recently published findings from 781 Garmin Vivosmart 4 users, collected over three months, reporting no correlation between the stress levels shown on the watch and the self-reported feelings of stress by the participants.
Typically, this has been jumped on by the media, quick to write headlines like "Smartwatches offer little insight into stress levels" and "Your smartwatch is lying to you about stress levels" and so on. However, as someone whose job it is to test the best smartwatches and fitness trackers on the market, I think there's more to this than meets the eye.
At first glance, the research seems pretty damning: in the three-month window, participants were asked to regularly self-report how they felt they were sleeping, along with information about their perceived tiredness and stress levels.
The study, published in the Journal of Psychopathology and Clinical Science, had this to say regarding the results:
"Self-report and wearable measures of sleep-related variables showed robust associations, while associations were weaker for tiredness, and measures of stress did not overlap for most individuals. These findings suggest that wearable data and their corresponding self-report measures may not necessarily measure similar constructs."
So while the Vivosmart 4 is good for sleep tracking and lines up with self-reported sleep quality, it struggles to tell when you're tired and stressed.
The researchers were aiming to see if wearable technology can provide a "warning system" for depression in students, using the above biomarkers. Most headlines have largely ignored this or buried the lede, conflating stress with excitement and high heart rate.
The stress Garmin watches refer to is physiological stress, which happens to your nervous system when your heart beats faster and you start to sweat. Exercise, excitement, arousal, and other emotional factors can cause this to happen to your body. It's impossible for LED heart rate and body temperature sensors to read your mood, so it's all logged as 'stress'.
I'm currently testing the WHOOP MG, and its app also contains a stress metric. On asking its AI coach – what a time to be alive – whether stress meant emotional or physiological, my app had this to say:
"The Stress metric in the WHOOP app is physiological – it's based on signals from your body, not how you feel emotionally."
"While emotional stress (like anxiety or pressure at work) can show up in your score, so can things like a tough workout, poor sleep, or even fighting off a cold. The Stress Score is a window into how your body is handling all types of stressors, not just your mood."
(Image credit: Future)It's easy to see how people get confused, and there is a grain of truth in that the headlines would be correct in assuming smartwatches aren't very good at calculating people's moods.
However, that's because they weren't designed to actually record mood, rather than any failing on their part. The whole study was testing whether smartwatches could be used as a depression predictor, and while indicators such as lack of sleep can be accurately measured, they can't do moods – especially if people are recording themselves as feeling stressed out, rather than reporting physiological sources of stress.
My tip? Use the stress metric in tandem with other metrics, such as sleep and calories burned, to understand when it's time to take a rest if you're able to do so. If your stress is high and your sleep is low, but you feel otherwise happy, you might want to skip the bars or the gym this Friday night and have a relaxing evening with a book to bring that stress score down.
You might also likeGoogle has announced plans to invest $9 billion to expand its cloud and AI infrastructure in Oklahoma over the next two years.
As part of the multi-billion dollar investment, the company has committed to building a new data center campus in Stillwater and expanding its existing Pryor data center, which opened in 2011 and has received an estimated $4.4 billion to date.
Oklahoma citizens will also benefit from a separate pot of funding which Google plans on sharing across multiple US states to boost AI skills.
Google is spending billions on Oklahoma AI"As part of a broader $1 billion commitment to American education and competitiveness, the University of Oklahoma and Oklahoma State University are part of the first cohort of the Google AI for Education Accelerator," the company wrote.
The scheme will see participants access Google Career Certificates and other AI training courses.
Acknowledging data center campuses come with broader environmental and sustainability impacts, Google will also partner with the electrical training ALLIANCE to boost the state's electrical workforce pipeline by 135% in anticipation of increased demand in the future.
Google President and Chief Investment Officer Ruth Porat announced the investment during a visit to the Pryor campus on August 13, 2025, noting, "Google is helping to power a new era of American innovation with our investments in Oklahoma."
"For nearly two decades, Google has proudly called Oklahoma home, starting with our data center in Mayes County, our second largest in the world," Porat added.
Since building out its first campus in the state, Google claims to have invested over $5.7 billion in the state, generating $2.2 billion of economic activity for state citizens and businesses in 2024 alone.
The company also matched 87% of its energy consumption with carbon-free energy in 2022-2023 from over 680MW of clean energy contracts in the state, boasting about other achievements like water replenishment and advanced power usage effectiveness in its impact study.
You might also likeNvidia has launched two new SFF (Small Form Factor ready) RTX Pro Servers and workstation GPUs built on its Blackwell architecture that use a single-fan, blower-style, low-profile design.
The new RTX Pro 4000 SFF and RTX Pro 2000 slot in beneath the top-end RTX Pro 6000 but are designed to maintain strong performance for professional workloads in smaller, lower-power workstation builds.
Both GPUs aim to shift more processing from CPU to GPU, allowing even small-form-factor workstations to handle AI inference, rendering, and simulation faster than before.
Compact hardware with workstation-level capabilityThe RTX Pro 4000 SFF is built with 24GB of ECC GDDR7 memory and delivers up to 770 AI TOPS.
Nvidia says it offers more than twice the AI performance of the previous-generation RTX A4000 SFF, and also features improved ray tracing, 50% more memory bandwidth, and has a 70W thermal limit, making it suitable for workstations with limited cooling capacity.
The RTX Pro 2000, also limited to 70W, carries 16GB of ECC GDDR7 memory and 545 AI TOPS of throughput.
Nvidia claims it is about 1.5 times faster than the RTX A2000 in 3D modeling, CAD, and rendering.
The company also points to its usefulness in AI-powered image and text generation, potentially speeding up design and content workflows in smaller studios or engineering teams.
These workstation GPUs share the Blackwell architecture benefits found in Nvidia’s larger RTX PRO lineup, including support for FP4 precision through fifth-generation Tensor Cores and the second-generation Transformer Engine.
This is intended to improve inference performance while keeping power consumption in check.
Nvidia has not yet disclosed specific pricing for these models but has indicated that it will make them available through partners such as PNY, TD Synnex, Dell, HP, and Lenovo later this year.
Given their position below the RTX Pro 6000, they are expected to cost considerably less while still providing a large performance increase over earlier small-format workstation GPUs.
While marketed as an upgrade path for professionals, the actual advantage over the fastest CPU-only small workstation configurations will depend on workload type and software optimization for GPU acceleration.
For tasks that already benefit heavily from GPU compute, such as AI inference, large-scale rendering, and simulation, the new RTX Pro 4000 SFF and RTX Pro 2000 could offer strong gains without requiring a larger, more expensive workstation.
For CPU-focused applications, the improvements may be less dramatic, but the added GPU capability still expands the performance envelope for small, power-limited systems.
You might also likeResearchers at Bitdefender have announced two critical vulnerabilities affecting a large number of Dahua smart cameras.
The flaws, which were patched in the most recent firmware update, could allow unauthenticated attackers to take full control of affected devices.
Dahua has confirmed that a total of 126 models were affected, including multiple IPC, SD, and DH series devices, not just the Hero C1 model first reported.
Patch nowThe first of the vulnerabilities, CVE-2025-31700, is a buffer overflow flaw in Dahua camera firmware that can be triggered when the device processes specially crafted network packets. If exploited, it could cause the camera to crash or, in some cases, allow a remote attacker to run their own code on the device.
The second, CVE-2025-31701, is another buffer overflow issue also exploitable through maliciously crafted packets sent over the network. It too can be used to crash the camera or potentially gain full remote control depending on the target’s defenses.
Both can be exploited to run arbitrary code with root privileges.
Bitdefender privately reported the issues to Dahua on March 28, 2025. The Chinese video surveillance equipment manufacturer acknowledged the report the next day and validated the findings by April 1.
It requested some time to prepare a fix for the issues, with patches finally rolling out last month, followed by the agreed public disclosure.
The two vulnerabilities can be especially dangerous for devices accessible from the internet via port forwarding or UPnP, as no authentication is required for possible exploitation.
Bitdefender warns that successful attacks could bypass firmware integrity checks and deploy persistent malicious code, making cleanup difficult.
Dahua, the world’s second-largest CCTV manufacturer behind Hikvision, has faced scrutiny in several countries over cybersecurity issues and data privacy concerns, particularly related to potential vulnerabilities in its network-connected devices.
It maintains a Product Security Incident Response Team (PSIRT) to coordinate with researchers on reported flaws, such as in the case of these vulnerability disclosures.
It is urging all customers who have not yet done so to update their camera firmware as a matter of urgency.
For anyone unable to do so immediately, it advises disconnecting vulnerable devices from direct internet access, disabling UPnP, and isolating cameras on separate networks to reduce risk.
A detailed list of affected models is included in Dahua’s online advisory, along with links to patched firmware.
Both Dahua and Bitdefender stress unpatched internet-connected devices should be considered prime targets.
You might also likeLG and Panasonic are just two of the top brands in the OLED TV market, but unlike rivals such as Samsung, Sony and Philips, their TVs often share a lot in common, right down to using the same OLED display panels.
The LG G4 and Panasonic Z95A, two of the best OLED TVs launched in 2024, both used micro-lens-array (MLA) OLED panels manufactured by LG. In 2025, it’s a similar story as the LG G5 and Panasonic Z95B both use LG’s new Primary Tandem RGB OLED panel (also known as four-stack OLED).
That means you should expect a similar picture from both, right?
As I’ve discovered when testing combinations of some of the best TVs, models that use the same display panel don’t always look the same. For example, the Samsung S95F and Sony Bravia 8 II flagship OLED TVs use Samsung’s QD-OLED panels, but when I tested them side by side, I got very different results.
The LG G5 and Panasonic Z95B were featured in our recent OLED four-way showdown with a casual judging panel, with the aforementioned S95F and Bravia 8 II rounding out the group, but I thought it would be illuminating to compare the LG and Panasonic on their own.
Brightness Image 1 of 2In default Filmmaker Mode, the LG G5 (right) clearly has the brightness advantage over the Panasonic Z95B (right) (Image credit: Future)Image 2 of 2But adjust the Z95B's brightness, and it gets much closer to the G5 (Image credit: Future)With both the Z95B and G5's default Filmmaker Mode settings active, it was clear that the G5 had a huge brightness advantage – an odd result considering both use the same OLED panel. When viewing some demo footage from the Spears & Munsil UHD Benchmark 4K Blu-ray in HDR10 format, scenes with snow looked far brighter and more vibrant on the G5, with much more brilliant whites and highlights. The Z95B’s picture still looked very accurate, but dim in comparison.
Leaving both sets in Filmmaker Mode, I measured the Z95B’s peak HDR brightness on a 10% white window pattern at 1,028 nits, a significantly lower result than the G5’s 2,268 nits on the same pattern.
After a quick check of the Z95B’s picture settings, I found its Luminance Level (brightness) was set to 70/100 by default, while the G5’s was set to 100/100. After upping the Z95B’s brightness to 100, it hit 2,355 nits, making it the brightest OLED I’d measured to date on this test.
Going back to the same Spears & Munsil footage with the brightness on the Z95B set to 100, the snow scenes now showed much more dazzling whites. It looked very similar to the G5, with the only real difference being the color temperature. I did, however, feel that despite the boost in brightness, the Z95B lost a little of the picture accuracy I’d noted previously.
Color and contrast With Dolby Vision sources, such as Wicked (pictured) the Panasonic Z95B (left) and LG G5 (right) have very similar colors. (Image credit: Universal Pictures / Future )Both the Z95B and G5 had superb color in my comparison. Watching more Spears & Munsil footage in HDR10 format, a parrot’s yellow and green feathers looked vivid on both TVs, and a field of red flowers was crisp and refined, with plenty of punch. Once again, the G5’s brightness in default settings gave colors more pop, but upping the Z95B’s brightness brought it to a similar level of vibrancy. The Z95B demonstrated richer, deeper colors, though, thanks to its stronger contrast and more refined black levels.
Switching to Wicked on 4K Blu-ray in Dolby Vision, the default brightness settings were 100/100 in both the Dolby Vision Dark mode on the Z95B and Dolby Vision Filmmaker Mode on the G5. This made the colors on both TVs look very similar. Elphaba’s green skin and the pink flowers and blue details on a wall in the Wizard & I scene both had the same eye-popping, vibrant color, and they also looked true-to-life.
Where the TVs differed was that the Z95B's deeper black levels made colors appear bolder and more detailed, whereas they looked brighter and punchier on the G5. This was easy to see in Glinda’s pink outfits and the greens of the Emerald City in Wicked.
Both the Panasonic Z95B (left) and LG G5 (right) deliver excellent contrast and black levels, but the Z95B looked more accurate on The Batman (Image credit: Warner Bros. / Future)As you’d expect from two top-tier OLED TVs, both black levels and contrast are excellent. Watching Alien: Romulus in Dolby Vision on 4K Blu-ray, shots of space or dark tunnels within the ship showcased rich black levels on both TVs, with excellent contrast between dark shadows and bright highlights from stars, lights and torches. The Z95B had the edge of the two TVs here with its deeper blacks, but again, both looked great.
Switching to The Batman in Dolby Vision on 4K Blu-ray, I used the opening crime scene section I regularly use for testing contrast. For this movie, I had to view in dimmed or pitch black conditions as both the Z95B and G5 struggled with the overhead lighting in our testing lab.
Interestingly, although both showcased excellent black levels and contrast, with the torches carried by detectives balancing well against the dim surroundings, I noticed that in pitch black conditions, black levels were more elevated on the G5. The Z95B’s deeper blacks resulted in stronger contrast, although there was some loss of shadow detail, specifically Batman’s eye and the logo on his chest when he looked towards the camera. Of the two, I found myself drawn to the Z95B as it felt more accurate to the movie.
Bang for your buck
Both the Z95B (left) and G5 (right) are brilliant OLED TVs, but your choice may come down to personal preference and price. (Image credit: Future)It’s fair to say that in recent years, a drawback of Panasonic’s flagship OLED TVs has been their price, which is often hundreds more than that of its main rivals at launch. This year, however, Panasonic is being more aggressive with its pricing. The 55-inch Z95B is available for $2,399 / £2,299, while the 55-inch G5 is available for $2,199 / £2,299. The 65-inch Z95B, meanwhile, is available for $3,099 / £2799, while the 65-inch G5 is $2,899 / £2.899.
One thing that could swing things in the Z95B’s favor between these two TVs is sound. Panasonic’s TVs are always among the best TVs for sound, and the Z95B is no different.
Watching the Batmobile chase scene from The Batman, the Z95B stood head and shoulders above the G5 when it came to the built-in sound. The Z95B’s punchier sound and much more powerful bass better captured the rumble of the Batmobile’s engine and the crunching of car impacts and explosions. Speech was clear on both sets, but the Z95B delivered better clarity. With the Z95B, you can easily watch without using one of the best soundbars – a factor that saves you money – whereas a soundbar is more recommended with the G5.
Where the G5 has the edge is in its gaming features, which are up there with the best gaming TVs and include 4K 165Hz support on four HDMI 2.1 ports. The LG’s webOS 25 smart TV interface is also one of the best I’ve used and provides a superior overall experience to the Z95B’s Fire TV smart interface.
Both of these are class-leading OLED TVs, and both are in the conversation for TV of the year. But based on this comparison, I found myself favoring the Panasonic Z95B. Its default Filmmaker Mode brightness setting may be odd, but once adjusted, I found myself more drawn to its picture than the LG G5's picture.
You might also likeAttackers are increasingly personalizing phishing emails to deliver malware, experts have warned, with criminals reaping in huge gains.
In adding the recipient’s name, company and other details into subject lines, file names, and message content, threat actors seek to make the messages appear more legitimate, increasing the chances that recipients will open malicious attachments or click links, researchers at Cofense have revealed.
Cofense analyzed a year’s worth of data and found that while several campaign themes use this tactic, finance-themed phishing was the most worrying due to both its frequency and impact.
Stay safeNearly 22% of subject-redacted emails fell into this category, often posing as invoices, tenders, or payment summaries.
Many of these emails carried jRAT, a cross-platform remote access trojan that can give attackers full control of a system, steal files, and install more malware.
Finance-themed phishing is particularly effective because it blends seamlessly with normal workplace communication, as employees often expect emails about contracts or payment updates.
While finance-themed phishing accounted for 21.9% of personalized subject cases, other themes also made heavy use of this approach.
Travel Assistance was the largest category at 36.78%, often used to deliver Vidar Stealer under the guise of reservation or itinerary updates.
Response-themed emails followed at 30.58%, frequently carrying PikaBot in messages disguised as meeting cancellations or order confirmations.
Tax-themed campaigns made up 3.72%, commonly involving Remcos RAT in password-protected archives, while Notification-themed phishing also represented 3.72%, delivering various malware families including WSH RAT and jRAT.
To counter these threats, Cofense advises verifying unexpected email requests through trusted channels, keeping antivirus and malware removal tools up to date, and limiting public exposure of staff details to make targeting harder.
Summing up, Cofense says, “While customized subject lines are not used in all malware email samples, it is a strong tactic to make the recipient feel a higher sense of urgency that may lead to a successful infection. Particularly targeted emails delivering RATs or Information Stealers can be notable for potentially providing remote access or login credentials that can be brokered to ransomware threat actors.”
You might also likeGeoffrey Hinton, scientist, former Google employee, and widely recognized 'Godfather of AI,' has made a late-stage career of criticizing his godchildren. And now he's taken it all a step further, insisting we need "AI Mothers," not AI Assistants.
Speaking at the AI4 Conference in Las Vegas this week, and as first reported by Forbes, Hinton again sounded the alarm on the impending advent of Artificial General Intelligence, which he now believes will arrive in a few years, a notion that syncs with recent comments from OpenAI CEO Sam Altman.
That acceleration from what was once thought to be decades to a few orbits around the sun is, perhaps, what prompted Hinton to argue that we need something other than AI Assistants.
"We need AI mothers rather than AI assistants," Hinton said, according to Forbes. The idea, Hinton posits, is that AI's with "maternal instincts" are a sort of protection system. After all, mothers generally don't harm and usually protect their children.
If AI systems like ChatGPT, Claude AI, and Gemini truly become smarter than us in a matter of years, having them in some way feel as if it's their job to look out for us might prevent them from harming us or society.
Hinton, who recently won a Nobel Prize and helped develop the technological foundation that arguably made all this AI possible, left Google in 2023 and immediately started warning people about a dire AI future. Imagine a parent disowning their child, and you get the idea.
I don't think Hinton is turned off from AI. After all, he can't stop talking about it, and appears to recognize its potential, but it's also clear it scares him.
He previously told The New York Times in 2023 that
So, sure that day is now fast approaching, but is a motherly AI what we want or need? I don't think so.
The minute we start training "Mom Instincts" into AI, it will start to act like a mother and slip into that creepy, uncanny valley where you can no longer tell if you're talking to a program or a person. Motherly instincts imply warmth, compassion, caring, understanding, and love. I don't want those things from an AI.
What I think we need, though, is for AI assistants to understand what it means to be human. Put another way, if AI chatbots can at least understand humanity, they can serve us better. They can also recognize our propensity for trust and perhaps finally stop presenting us with false narratives and fake friendliness and interest.
We shouldn't want companionship out of our super-intelligent AI systems. Instead, we need utility and trust, an ability to carry out our wishes in a way that best serves our interests.
The last thing we need is an AI full of maternal instincts, which then makes its own choices and, when things go awry, insists, "Well, dear, mother knows best."
You might also likeComedy Central did not air South Park season 27 episode 3 on its expected release date (August 13), with the series now thought to continue next week instead. This also means it now won’t be available on Paramount+ either. However, unlike the season’s previous delay, the episode 3 delay was scheduled in advance. Instead, the network declared the day to be “South Park day,” running a marathon of fan-favorite episodes followed by the sitcom’s pilot episode ‘Cartman Gets an Anal Probe’ in the typical 10pm ET timeslot.
Episode 2 was delayed after the show’s controversial premiere, which featured the likes of a deepfaked Donald Trump with no clothes on and ICE (Immigration and Customs Enforcement) agents shooting puppies and driving around ominously in large groups. Screenshots from the show were quickly used on both the White House and US Secretary of Homeland Security Kristi Noem’s social accounts in order to try and recruit more people to join ICE.
The Department of Homeland Security (NHS) additionally told Newsweek: “We want to thank South Park for drawing attention to ICE law enforcement recruitment. We are calling on patriotic Americans to help us remove murderers, gang members, pedophiles, and other violent criminals from our country.”
It’s for this reason that I can’t help but think the delay of South Park season 27 episode 3 has something more attached to it. It’s certainly a conveniently timed break in the schedule, but given that season 27 has only been on for three episodes yet has suffered two delays already. Surely things aren’t operating smoothly behind the scenes? We always know what we get into with the hit TV show, but that might no longer be to everybody’s tastes.
Has South Park season 27 hit a political snag after its season 3 delay? I wouldn’t be surprisedKristi Noem's parody in South Park season 27. (Image credit: Paramount)I don’t need to spell out that the political and cultural landscape has remarkably changed in the last few years, and that’s possibly to South Park’s detriment. While real-world political and cultural affairs are offering a smorgasbord of inspiration, it also comes with a more critical lens. We’ve already had the comments from the NHS, above, with Kristi Noem also weighing in on her portrayal as well. Before her secret post-credits scene was revealed, she responded during an interview on the Glenn Beck Program podcast, slamming her character’s appearance: “It’s so lazy to just constantly make fun of women for how they look. If they wanted to criticize my job, go ahead and do that. But clearly they can’t – they just pick something petty like that.”
White House spokesperson Taylor Rogers also told Variety about the events of episode 1: "Just like the creators of South Park, the Left has no authentic or original content, which is why their popularity continues to hit record lows. This show hasn't been relevant for over 20 years and is hanging on by a thread with uninspired ideas in a desperate attempt for attention."
This won’t be the last round of comments from political figures about what’s happening on the show, and I wouldn’t be surprised if we see public conversations between the two sides on a weekly basis. I also wouldn’t be surprised if there’s resistance to what’s being shown behind the scenes, given nobody exactly comes across particularly well if they’re being featured on South Park. We’ve seen critics of the current US Government coming under fire or having their shows canceled (The Late Show with Stephen Colbert’s cancellation is a great example of this), so discussions about what should and shouldn’t be parodied could be holding up episode releases.
Of course, this is all speculation. We’ve got no idea what the rest of South Park season 27’s 10-episode run has in store for us, and I really hope Paramount and the show’s creators can hold firm to delivering the creative vision they want to. I can’t deny that I’m worried about episodes 4-10, but I’m also holding out hope. After all, South Park wouldn’t be South Park if it wasn’t near the knuckle and beyond, right?
You might also like