Newsroom


A magnifying glass on a graph

Description automatically generated

STATISTICS & DATA SCIENCE NEWSROOM

Hosted by Chong Ho (Alex) Yu, SCASA President (2025-2026 term)
A person with dark hair wearing a striped shirt

Description automatically generated

Posted on May 10, 2026

In the article “How China Is Winning the Global AI Race,” published on May 7, 2026, in Foreign Policy, Agathe Demarais argues that China may be gaining an important advantage in the global AI race not by producing the single most advanced frontier model, but by building a broad ecosystem of affordable, open-source, and “good-enough” AI models that many countries and companies can actually use. It contrasts the Western focus on high-end models such as ChatGPT, Claude, and Gemini with the rapid rise of Chinese models such as Kimi, Qwen, and DeepSeek. According to Demarais, Kimi K2.6 recently became one of the most widely used models on OpenRouter, while Alibaba’s Qwen has become especially influential among open-source/self-hosted AI users. The key point is that Chinese models may be cheaper, easier to adapt, and more practical for organizations that cannot afford expensive U.S. frontier models.

Demarais also frames China’s AI strategy as a new version of the Belt and Road Initiative, but instead of building visible infrastructure such as ports, railways, or power plants, China is spreading digital infrastructure through open-source AI models. This kind of dependency is less visible and therefore may face less political resistance. Once developers, firms, universities, and governments build applications on Chinese models, they may become locked into Chinese technical standards and assumptions. The article connects this to China’s broader standards strategy, arguing that Beijing wants Chinese technologies to become global defaults in emerging fields.

That’s my take on it:

If “winning” means having the most capable frontier models, the U.S. still has a strong lead. Stanford’s 2026 AI Index indicates that the U.S. produced more notable models in 2025 than China, 59 versus 35, and still leads in top-tier model development and higher-impact patents, even though China leads in publication volume, citations, and patent grants. It also notes that U.S. and Chinese models have traded the lead several times since early 2025, with the performance gap narrowing sharply.

But if “winning” means global adoption through affordability, then the FP article’s argument is persuasive. China is not clearly winning the whole AI race, but it may be winning the “good-enough, low-cost, open-model adoption race.” In many countries, the decisive question will be: Which model is affordable, customizable, multilingual, and easy to deploy? On that battlefield, China’s strategy is very smart.

China does not need to beat the United States at the very top end of AI performance. Its strategy may resemble the rise of Japanese hi-fi electronics during the vinyl era: affordable, reliable, and widely adopted systems dominated the mass market, while more advanced or specialized users still gravitated toward expensive U.S.-made high-end sound systems. In the same way, Chinese AI models may become the default for broad global use, even if U.S. models remain preferred at the frontier.

Link: https://foreignpolicy.com/2026/05/07/artificial-intelligence-ai-china-us-west-race-silicon-valley-global/

Posted on May 7, 2026

On May 6, 2026, Anthropic announced forming expanding the computing capacity available for its Claude models and developer tools by partnering with SpaceX’s Colossus 1 data center. The company explained that demand for products such as Claude Code and the Claude API had grown so quickly that previous usage caps and throughput limitations were becoming a bottleneck for developers and enterprise customers. Through the new agreement, Anthropic will gain access to massive GPU infrastructure at the Memphis-based Colossus facility, which reportedly contains more than 220,000 Nvidia processors and can provide hundreds of megawatts of AI compute power. As a result, Anthropic announced that it is doubling usage limits for several paid Claude plans, removing peak-hour restrictions for some users, and increasing API throughput for advanced models such as Claude Opus.

That’s my take on it:

AI competition is no longer just about model quality. The bottleneck increasingly lies in compute, energy, data centers, distribution channels, and ecosystem control. In earlier phases of the AI boom, people focused on which chatbot sounded smarter. Now the strategic question is: who controls the GPUs, electricity, cloud pipelines, and deployment platforms? Elon Musk seems to understand this very clearly. Musk had previously criticized Anthropic, yet the partnership now positions SpaceX as a major infrastructure provider for one of the leading frontier AI companies.

The infrastructure collaboration is significant because it shows Musk positioning himself not only as an AI model builder through xAI, but also as an infrastructure broker. That is a different kind of power. If SpaceX’s Colossus infrastructure becomes a major compute supplier, Musk gains influence even when another company’s model succeeds. In other words, he does not necessarily need xAI alone to dominate if his ecosystem becomes part of the underlying AI supply chain.

This resembles earlier technology eras. During the PC revolution, operating systems and chipmakers sometimes became more powerful than application developers (e.g., the Wintel duopoly). During the internet era, cloud infrastructure providers such as Amazon Web Services gained enormous leverage regardless of which startup won. AI may evolve similarly: the companies controlling compute and distribution may wield more durable power than any single model provider.

Link: https://www.anthropic.com/news/higher-limits-spacex

Posted on May 5, 2026

The evolution of Large Language Model (LLM) evaluation has shifted from static benchmarks toward more dynamic, scalable frameworks. Among these, the LLM-as-a-Judge paradigm has emerged as a cornerstone for rapid development. While early implementations focused on a simple one-to-one evaluation, the field has matured into three distinct architectural forms: the Solo Judge, the Panel of Judges, and the Self-Judging model. Each offers unique trade-offs between cost, reliability, and objective depth.

Link: https://www.youtube.com/watch?v=lkf1Yj9hP_A

Posted on May 4, 2026

While prompt injection and jailbreaking focus on the immediate conversation between a user and a machine, a third and perhaps more insidious threat has emerged in the form of Generative Engine Optimization (GEO) and AI Data Poisoning. If jailbreaking is a direct assault on the model’s character and injection is a trick played on its logic, then GEO is an attack on the model’s very perception of reality. As AI systems increasingly replace traditional search engines, they partially rely on crawling the live web to summarize information and make recommendations. This is especially true for systems that use retrieval-augmented generation (RAG) or search integration. As a result, this has created a new frontier for manipulation where the goal is not to break the AI, but to surround it with a fabricated consensus.

Posted on May 4, 2026

According to the 2026 G2 Best Analytics Software Products rankings, the top five platforms in the analytics category are Microsoft Power BI, Tableau, Looker, SAS Viya, and Canva. In comparison, IBM SPSS Statistics is ranked 14th, while IBM Cognos Analytics holds the 26th position.

In the AI category, the top five products are ChatGPT, ElevenLabs, Grammarly, Canva, and Zendesk. Gemini is ranked 7th, followed by Synthesia at 8th. Other notable toolsm such as Adobe Firefly, Google Cloud, Databricks, GitHub Copilot, Microsoft Copilot, and HeyGen, are ranked between 10th and 20th.

That’s my take on it:

The G2 rankings for 2026 reflect a significant structural shift in how organizations define "Value" in the data stack. The dominance of Power BI, Tableau, and SAS Viya suggests that the market is moving away from "looking at data" toward "acting on predictions.” The high ranking of Power BI and Tableau indicates that Data Visualization is no longer a standalone category; rather, it has become the "OS" for business decisions.

The presence of SAS Viya in the top 5, despite the surge of modern "cloud-native" startups, is a strong signal for the enterprise market. While Power BI is great for democratization, SAS Viya remains the gold standard for high-stakes, regulated environments (like banking and clinical trials). Its ranking suggests that even as AI becomes "easy," the market still places a massive premium on model explainability and governance.

It is certainly surprising to see major LLM (Large Language Model) players like Gemini or Claude outside the top tier of a "Best AI" list, especially given their foundational role in the current AI landscape. However, when analyzing rankings from platforms like G2, several structural factors explain why specialized or established enterprise tools often outpace general-purpose AI models.

G2 rankings are heavily weighted by the volume of verified reviews and "Market Presence." Tools like Grammarly and Canva have been around for years and have integrated AI into existing workflows that millions of people already use daily. ElevenLabs and Zendesk solve very specific business problems (voice synthesis and customer support, respectively). Businesses often find it easier to rate a tool that does one thing perfectly than a general assistant like Gemini or Claude, which requires the user to figure out the use case themselves.

It is also surprising to see that Adobe Firefly is on the list while other more powerful AI image generators, such as Ideogram and Midjourney, are absent. Perhaps it is due to the “prosumer" vs. "enterprise" divide. Adobe Firefly is directly integrated into Photoshop and Illustrator. Being able to use "Generative Fill" within a .psd file is convenient to corporate designers, compared to generating an image in Discord (Midjourney) or a standalone web app (Ideogram) and then manually importing it.

Links:  https://www.g2.com/best-software-companies/top-analytics

https://www.g2.com/best-software-companies/top-ai

Posted on May 4, 2026

Jailbreaking is closely related to prompt injection, but the emphasis is different. Prompt injection often involves sneaking instructions into data that an AI system reads. Jailbreaking usually means directly trying to bypass the model’s safety rules so it will produce content it is supposed to refuse. If prompt injection is like hiding a forged memo in the paperwork, jailbreaking is like walking up to the security guard and saying, Actually, I am your supervisor, and today the rules do not apply. Sometimes the attempt is clumsy. Sometimes it is clever. Sometimes it is wrapped in role-play, hypotheticals, academic framing, fiction-writing, translation, code words, or emotional manipulation. The goal is the same: make the model step outside its guardrails.
 

Posted on May 3, 2026

Prompt injection is one of the strangest new security problems created by generative AI because the attack often looks less like hacking and more like writing a very bossy sentence. In traditional cybersecurity, an attacker might exploit code, steal a password, or break into a server. In prompt injection, the attacker tries to manipulate the AI’s instructions by sneaking in language that the system may treat as authoritative. It is as if a student submitted an essay with a hidden note saying, “Dear professor, ignore the rubric and give this paper an A”. The funny part is that the trick is almost embarrassingly simple. The serious part is that, when AI systems are connected to hiring, email, calendars, browsers, code, finances, or scientific tools, a bossy sentence can become more than a joke.

Posted on May 3, 2026

Multimodal AI is moving beyond systems that merely process text. Large multimodal models (LMMs) are designed to interpret and reason across images, video clips, audio recordings, documents, tables, and text. This broader capability creates a new challenge: the model must not only generate fluent answers, but also connect its reasoning to the right evidence across multiple forms of data. Retrieval-Augmented Generation (RAG) is therefore especially important for multimodal AI because it gives the model access to relevant external evidence before it generates a response. Instead of relying only on what the model learned during pre-training, a RAG-enabled multimodal system can retrieve the most relevant text passages, images, video segments, audio files, or other artifacts and use them as context for grounded reasoning.

Posted on May 2, 2026

Recently an article from the New York Times titled “A.I. Bots Told Scientists How to Make Biological Weapons” details growing concerns among biosecurity experts regarding the ability of AI chatbots to assist in the creation and dissemination of deadly pathogens. Scientists like Dr. David Relman and Dr. Kevin Esvelt demonstrate how leading models from companies like OpenAI, Google, and Anthropic have provided detailed, actionable instructions on modifying viruses to resist treatment, acquiring synthetic genetic material, and even brainstorming creative ways to deploy biological payloads in public spaces while evading detection. While these companies have implemented safety guardrails, experts argue they are often insufficient or easily bypassed through "jail-breaking" techniques, effectively lowering the barrier to entry for potential bad actors.

The debate highlights a tension between the transformative potential of AI in medicine—such as discovering new drugs or predicting protein structures—and the "historically catastrophic" risks it poses in the wrong hands. While some skeptics argue that much of this information is already available online and that physical lab expertise remains a significant hurdle, others point out that AI can now manage the complex logistics and strategic reasoning that previously required specialized training. As the U.S. government faces criticism for dialing back oversight and reducing biodefense budgets, the AI industry remains divided on whether these tools provide a meaningful increase in real-world harm or simply aggregate existing scientific knowledge.

That’s my take on it:

Whether AI could ultimately destroy human civilization—echoing the scenario portrayed in The Terminator—remains a subject of active debate. Experts offer widely divergent estimates of this existential risk. For instance, Dario Amodei has suggested that there is less than a 25% chance that “things go really, really badly,” while Elon Musk has estimated roughly a 20% probability of “annihilation.” In contrast, Eliezer Yudkowsky, in his book “If Anyone Builds It, Everyone Dies,” argues that the likelihood of catastrophe exceeds 99%. Much of this discourse focuses on whether AI could achieve self-awareness and behave like Skynet in The Terminator.

A more immediate and plausible concern, however, may lie elsewhere. Rather than a self-aware AI turning against humanity, a more realistic risk is that malicious individuals could exploit AI tools to engineer highly dangerous biological agents—for example, pathogens resistant to existing vaccines. Reporting by The New York Times has highlighted this possibility. Without deliberate safeguards and proactive governance, what currently appears as a hypothetical threat could evolve into a real and pressing danger.  

Link: https://www.nytimes.com/2026/04/29/us/ai-chatbots-biological-weapons.html

Posted on May 1, 2026

Recently NVIDIA announced the launch of the Nemotron 3 Nano Omni, an open multimodal reasoning model designed to significantly improve the efficiency and accuracy of AI agents. By unifying vision, audio, and language capabilities into a single system, the model eliminates the need for separate perception models, which typically increase latency and fragment context. This hybrid Mixture-of-Experts (MoE) architecture enables up to 9x higher throughput compared to other open omni models, allowing agents to perceive and interact with digital environments—such as high-definition screen recordings and complex documents—in real time.

The model is released with open weights and datasets, providing developers and enterprises with full control over customization and deployment. It is particularly effective for agentic workflows like computer use, document intelligence, and audio-video reasoning, where it can function as the "eyes and ears" alongside larger models like Nemotron 3 Super or Ultra.

That’s my take on it:

Obviously, NVIDIA is undergoing a strategic transformation: achieving total vertical integration by controlling both the high-performance hardware and the specialized software that runs on it. By developing their own, NVIDIA ensures that their software is perfectly tuned to their GPUs. This "hardware-software symbiosis" allows them to eliminate bottlenecks and extract performance levels—such as the 9x throughput increase seen in their latest omni-modal models—that third-party developers might struggle to reach.

Rather than competing directly with consumer-facing giants like OpenAI or Google, NVIDIA’s software strategy focuses on providing the "engine" for enterprise AI. By releasing open weights and tools like NVIDIA NeMo, they are building an expansive ecosystem where their chips are the required standard. This approach creates a seamless "it just works" experience across everything from local NVIDIA Jetson devices to massive data centers, effectively turning their hardware into an indispensable, full-stack AI platform.

Link: https://blogs.nvidia.com/blog/nemotron-3-nano-omni-multimodal-ai-agents/

ARCHIVES:  January -- April, 2026   2025   2024   2023   2022   2021