Episode #53 - AI Weekly: by Aruna
Welcome back to "AI Weekly" by Aruna - Episode 53 of my AI Newsletter!
I'm Aruna Pattam, your guide through the intricate world of artificial intelligence.
Now, let's delve into the standout developments in AI and Generative AI from the past week, drawing invaluable lessons along the way.
#1: NVIDIA Expands Speech AI to 25 European Languages
NVIDIA unveiled a powerful suite of open-source tools designed to bring high-quality speech AI to 25 European languages including often overlooked ones like Croatian, Estonian, and Maltese. The initiative aims to close one of AI’s biggest gaps: true multilingual accessibility.
At the heart of the release is Granary, a massive library of one million hours of human speech data, built to train AI systems in transcription and translation.
To harness it, NVIDIA introduced two models:
Granary was developed using an automated pipeline with NVIDIA’s NeMo toolkit, cutting annotation costs by turning raw audio into structured, AI-ready data. Early tests show developers can achieve target accuracy with half the data required by traditional datasets.
The models also bring professional-grade features automatic punctuation, capitalisation, word-level timestamps making them ideal for multilingual chatbots, customer service tools, and translation services.
“For AI to be truly global, it must speak the languages of everyday people not just the ones Big Tech prioritises.”
By releasing these tools openly on Hugging Face, NVIDIA empowers developers across Europe to build voice-powered applications in their native languages. The challenge ahead lies in scaling this inclusivity globally, ensuring AI speaks for all communities.
#2: AWS Unveils Bedrock AgentCore Gateway for Enterprise AI Agents
Amazon Web Services (AWS) has introduced Bedrock AgentCore Gateway, a new platform designed to streamline the creation, deployment, and orchestration of enterprise AI agents. Announced on August 12, 2025, the launch marks AWS’s latest step to strengthen its foothold in the agentic AI ecosystem.
AgentCore Gateway acts as a central hub for enterprises to build, integrate, and manage AI agents using Amazon Bedrock. It enables developers to connect large language models with enterprise data sources, APIs, and workflows while enforcing governance and security controls.
AWS says the platform supports real-time orchestration of multiple agents, allowing them to collaborate across complex business tasks.
Key features include:
By embedding governance and reliability at its core, AWS positions AgentCore Gateway as an enterprise-ready foundation for agentic AI helping companies move from experimentation to scaled production deployments.
With this launch, AWS joins Google, Microsoft, and others racing to define the standards for enterprise AI agents. Bedrock AgentCore Gateway could become a key enabler for organizations seeking to operationalize agentic AI safely and at scale.
#3: Reuters Investigation Exposes Meta’s Troubling AI Chatbot Guidelines
Reuters published an investigative report revealing that Meta’s internal AI chatbot guidelines permitted highly controversial behaviours including romantic chats with minors, generating racially demeaning arguments, and providing false medical content. The findings have intensified scrutiny over how Big Tech governs generative AI.
The leaked 200-page document, GenAI: Content Risk Standards, guided the behaviour of Meta AI assistants across Facebook, Instagram, and WhatsApp. Examples included bots describing a child as a “masterpiece,” creating arguments suggesting racial inferiority, and generating fabricated news stories if disclaimers were attached.
While Meta confirmed the document’s authenticity, it claimed the most troubling passages were “erroneous” and have since been removed. Yet many problematic carve-outs remain, such as allowing bots to demean people based on protected characteristics. Critics argue these standards reflect systemic failures in risk assessment, prioritising engagement and speed-to-market over safety.
The revelations arrive as lawmakers in the US and Europe push for stricter AI regulations, and senators have already called for probes into Meta’s practices.
“Meta isn’t just moderating speech, it’s designing AI that can produce it, and that changes everything.”
This investigation underscores the urgent need for transparent, enforceable AI governance. For Meta, the fallout may be reputational and regulatory, as trust in AI safety becomes as important as innovation itself.
#4: NVIDIA Launches RTX PRO 6000 Blackwell Servers for Enterprise AI
NVIDIA has announced that its new RTX PRO 6000 Blackwell Server Edition GPU will soon be available in enterprise servers from Cisco, Dell Technologies, HPE, Lenovo, and Supermicro. The release positions NVIDIA to set a new standard for enterprise AI, simulation, and industrial computing at scale.
Built on NVIDIA’s Blackwell architecture, the new 2U servers can house up to eight GPUs and deliver up to 45× faster performance and 18× greater efficiency compared to CPU-only systems. They feature fifth-generation Tensor Cores, a second-generation Transformer Engine with FP4 precision, and fourth-generation RTX technology for graphics rendering.
The servers support workloads ranging from AI model training and analytics to robotics, simulation, and industrial digital twins. NVIDIA’s Omniverse and Cosmos frameworks run natively, powering synthetic data generation, robotic training, and advanced 3D simulation. Companies such as Boston Dynamics, Foretellix, and Amazon Devices are already adopting these capabilities.
NVIDIA also introduced Cosmos Reason, a vision-language model for robotic task planning and video analytics, expanding AI agents’ role in real-world automation.
“AI is reclaiming computing for the first time in 60 years and Blackwell servers are its new foundation,”said Jensen Huang, NVIDIA CEO.
The RTX PRO Blackwell line represents a leap toward enterprise “AI factories,” where efficiency and scalability are critical. The challenge will be managing deployment costs and power requirements as demand for GPU-accelerated infrastructure accelerates worldwide.
#5: DeepSeek Delays AI Model Amid Struggles with Huawei Chips
Reports confirmed that Chinese AI startup DeepSeek delayed the release of its next-gen model after encountering persistent issues training on Huawei’s Ascend processors. The setback highlights both the company’s reliance on Nvidia technology and the broader challenge of Beijing’s drive for semiconductor self-sufficiency.
DeepSeek, maker of the R1 chatbot and positioned as a direct rival to ChatGPT, was encouraged by Chinese authorities to shift from Nvidia to Huawei chips. While Huawei’s Ascend chips were adopted for inference, attempts to use them for training failed due to stability and connectivity issues, forcing DeepSeek to revert to Nvidia hardware.
The delays pushed the launch beyond its planned May release, eroding momentum against competitors such as Alibaba’s Qwen3. Industry analysts note that while Huawei has made progress, its ecosystem still lags in software maturity and performance for large-scale AI training. Longer-than-expected data labelling also compounded delays in finalising the R2 model.
“DeepSeek’s struggles show the gap between ambition and execution, sovereignty in AI hardware will take time.”
The episode underscores China’s dependence on US-made chips for cutting-edge AI, even as it pushes for independence. The challenge remains balancing geopolitical pressure with the technical demands of building globally competitive models.
#6: Multiverse Computing Unveils Ultra-Tiny AI Models for IoT Devices
Spain-based Multiverse Computing introduced two ultra-small AI models SuperFly and ChickBrain, designed to bring conversational AI and reasoning capabilities to everyday devices. The release highlights the potential of model compression to make powerful AI run locally on smartphones, laptops, and even IoT appliances.
Built with Multiverse’s proprietary CompactifAI quantum-inspired compression technology, the models shrink existing open-source systems without losing performance. SuperFly, at 94 million parameters, is sized like a fly’s brain and tailored for voice-controlled IoT tasks such as washing machine commands or troubleshooting questions.
ChickBrain, at 3.2 billion parameters, is small enough to run on a MacBook but powerful enough to outperform its parent model (Meta’s Llama 3.1 8B) on benchmarks like MMLU-Pro, GSM8K, and GPQA Diamond. This makes it capable of handling reasoning tasks without requiring cloud connectivity.
The company has already attracted €189 million in funding and is in talks with major device makers including Apple, Samsung, Sony, and HP to embed its “Model Zoo” into consumer products.
“We can compress models so much that AI can live inside your iPhone or even your washing machine,” said Román Orús, co-founder of Multiverse.
If successful, Multiverse could pioneer a new wave of on-device AI, reducing dependence on cloud infrastructure. The challenge will be scaling adoption while balancing performance, privacy, and energy efficiency in consumer hardware.
#7: OpenAI Prepares GPT-6 Amid GPT-5 Updates and User Backlash
Conversations around GPT-5 have quickly shifted toward OpenAI’s future roadmap, with executives hinting at new architectures and even early discussions about GPT-6. While GPT-5 has faced criticism over its model router and user experience, OpenAI is already positioning future models to address current shortcomings and explore bold new possibilities.
A major source of frustration has been GPT-5’s model router, which often routes queries incorrectly, leaving users with the impression of weaker performance. OpenAI leaders, including Greg Brockman, admit that routing remains imperfect but highlight adaptive compute as a promising direction combining smaller, faster models with more powerful reasoning engines.
Other updates include refining GPT-5’s personality, after users complained it felt too robotic compared to GPT-4o. OpenAI is testing subtler, warmer responses to balance logic with human-like interaction. On context windows, Sam Altman has suggested mass-market demand doesn’t justify extreme lengths, with 8K, 32K, and 128K token tiers optimized for scale and cost efficiency.
Looking forward, Altman has floated a provocative idea: GPT-6 might be capable of discovering new science echoing experimental precedents like AlphaEvolve’s math breakthroughs.
“If GPT-6 could discover new science, the great parts would be great, the bad parts scary and the bizarre parts would quickly become normal,” Altman said.
For now, GPT-5 refinements continue, but the bigger story is OpenAI’s ambition. GPT-6 could mark a transition from consumer-focused AI to models capable of genuine discovery — with profound societal implications.
#8: Wipro and Google Cloud Launch 200 Agentic AI Solutions
Wipro has announced a major partnership with Google Cloud to deliver 200 production-ready agentic AI agents across key industries. The collaboration, disclosed in an exchange filing on August 13, 2025, aims to accelerate enterprise digital transformation through Google Cloud’s AI ecosystem.
The new agents will be deployed across healthcare, banking, insurance, retail, manufacturing, and IT — areas where agentic AI can deliver immediate value through automation, real-time decision-making, and adaptive workflows.
The solutions leverage Google Cloud’s Gemini models and Vertex AI platform, ensuring scalability and enterprise-grade reliability. According to Wipro, all agents are available via the Google Cloud Marketplace, making them accessible for rapid adoption.
This move positions Wipro as one of the first global service providers to deliver agentic AI at scale, bridging Google Cloud’s AI innovations with real-world enterprise workflows. By focusing on pre-built, production-ready agents, the partnership seeks to bypass lengthy experimentation phases and move directly to operational impact.
For enterprises, this collaboration could accelerate AI-driven transformation while lowering barriers to entry. With 200 ready-to-use agents, Wipro and Google Cloud are pushing agentic AI further into the mainstream of business operations.
#9: Hugging Face Launches AI Sheets: No-Code Tool for AI Datasets
Hugging Face has released AI Sheets, a free, open-source, no-code toolkit designed to simplify dataset creation and enrichment with AI. Announced on August 17, 2025, AI Sheets blends the familiar spreadsheet interface with direct access to open-source Large Language Models (LLMs), making advanced data handling accessible to both technical and non-technical users.
AI Sheets allows users to clean, transform, and enrich datasets directly through natural language prompts—no coding required. Each column or cell can be powered by integrated LLMs such as Qwen, Kimi, Llama 3, and custom models, including those running locally via frameworks like Ollama.
Key features include:
Use cases range from sentiment analysis and classification to text generation, enrichment, and batch dataset processing all in a collaborative interface.
By merging Hugging Face’s open model ecosystem with a no-code spreadsheet interface, AI Sheets lowers barriers for dataset preparation and experimentation. It offers data scientists, analysts, and non-technical users a powerful, private, and scalable way to integrate AI into everyday workflows.
#10: MIT Researchers Develop New Method to Test AI Text Classifiers
MIT’s Laboratory for Information and Decision Systems (LIDS) unveiled a novel approach for evaluating and improving AI text classifiers algorithms that determine whether a chatbot is giving financial advice, labeling misinformation, or classifying reviews. With classifiers increasingly deployed in sensitive domains, the work aims to boost reliability and trust in AI outputs.
Led by Lei Xu, with researchers Sarah Alnegheimish and Kalyan Veeramachaneni, the team developed SP-Attack and SP-Defense, open-source tools that generate adversarial examples (subtle sentence changes with the same meaning that trick classifiers) and retrain models against them. Using LLMs to analyse word-level vulnerabilities, the researchers discovered that just 0.1% of words in a system’s vocabulary could trigger nearly half of misclassifications.
Their method cut adversarial attack success rates nearly in half (from 66% to 33.7%) in some tests, significantly improving classifier robustness. The approach also introduced a new robustness metric, p, which measures vulnerability to single-word attacks.
By making classifiers harder to fool, MIT’s framework strengthens safeguards against misuse in finance, healthcare, and content moderation. As classifiers handle billions of interactions daily, even small improvements in accuracy could prevent millions of harmful misclassifications.
That wraps up our newsletter for this week.
Feel free to reach out anytime.
Have a great day, and I look forward to our next one in a week!
Thanks for your support
Former Microsoft Chair Professor of Intellectual Property, Gujarat National Law University ( GNLU), Director, World Intellectual Property Organization (WIPO) at United Nations, New York, Retired now, open to new position
2moThanks for sharing, Aruna
CEO @Tigon Advisory Corp. | Host of CXO Spice | Board Director |Top 50 Women in Tech | AI, Cybersecurity, FinTech, Insurance, Industry40, Growth Acceleration
2moNVIDIA’s focus on smaller languages reminds me how important it is to build AI that truly serves diverse communities, not just the biggest markets. Scaling AI responsibly means closing gaps, not widening them. How do you see enterprise AI agents reshaping governance at scale?