AI Week # 14 Breakdown: Meta’s Llama 4 Controversy, NVIDIA’s Nemotron Leap & $1B+ in AI Funding

Week # 14 - Apr 2 to Apr 8

Last Updated on April 9, 2025 by Editor

Blink and you’ll miss the AI revolution! This week alone (April 2-8, 2025) unleashed a whirlwind of artificial intelligence breakthroughs you simply can’t afford to ignore. From Amazon’s hyper-realistic Nova Sonic voice AI and IBM’s powerhouse AI mainframe to Meta’s controversial Llama 4 models and Runway AI’s massive $308M funding haul for generative video, the pace is relentless. Get the essential updates on OpenAI’s strategic model releases and delays, Midjourney’s V7 image generator launch, pivotal chipmaking partnerships taking shape between Intel and TSMC, multi-million dollar AI funding rounds fueling innovation in sales and voice tech, and crucial cybersecurity investments – dive into AI Week 14’s highlights now, before the future leaves you behind!

Amazon Unveils Nova Sonic: A Leap in Real-Time Conversational AI

Amazon has introduced Amazon Nova Sonic, a cutting-edge speech-to-speech foundation model designed to enable remarkably natural, real-time voice conversations in AI applications. Integrated into Amazon Bedrock, Nova Sonic unifies speech understanding and generation into a single model, aiming to deliver human-like voice interactions with low latency and industry-leading price performance. It supports various English accents (American, British) and can adapt intonation and style based on context.

Nova Sonic boasts impressive performance, achieving a 4.2% word error rate (WER) on the Multilingual LibriSpeech benchmark, indicating 95.8% accuracy across multiple languages including English, French, Italian, German, and Spanish, even in noisy environments. It reportedly outperforms OpenAI’s GPT-4o by 46.7% on the Augmented Multi-Party Interaction benchmark for multi-participant conversations. With an average latency of just 1.09 seconds, it’s faster than GPT-4o’s Realtime API (1.18 seconds). Amazon highlights its cost-efficiency, claiming it’s around 80% more affordable than GPT-4o. Developers can leverage Nova Sonic via a new bidirectional streaming API on Bedrock for applications like customer service automation, voice assistants, and interactive learning.

Amazon’s Senior VP Rohit Prasad emphasized Nova Sonic’s role in advancing Artificial General Intelligence (AGI), with plans to expand its capabilities to modalities like image and video processing. Nova Sonic is poised to redefine voice-enabled applications across industries such as customer support automation, interactive education, and language learning.

Hire a Freelancer

I couldn’t access the specific webpages you linked. This sometimes happens due to paywalls, login requirements, or other access restrictions.

However, I was able to search the web for information on the topics mentioned in the URLs for the period around April 2nd to April 8th, 2025 (AI Week #14). Here’s a summary based on the information found:

Amazon Unveils Nova Sonic: A Leap in Real-Time Conversational AI

Amazon has introduced Amazon Nova Sonic, a cutting-edge speech-to-speech foundation model designed to enable remarkably natural, real-time voice conversations in AI applications. Integrated into Amazon Bedrock, Nova Sonic unifies speech understanding and generation into a single model, aiming to deliver human-like voice interactions with low latency and industry-leading price performance. It supports various English accents (American, British) and can adapt intonation and style based on context.

Nova Sonic boasts impressive performance, achieving a 4.2% word error rate (WER) on the Multilingual LibriSpeech benchmark, indicating 95.8% accuracy across multiple languages including English, French, Italian, German, and Spanish, even in noisy environments. It reportedly outperforms OpenAI’s GPT-4o by 46.7% on the Augmented Multi-Party Interaction benchmark for multi-participant conversations. With an average latency of just 1.09 seconds, it’s faster than GPT-4o’s Realtime API (1.18 seconds). Amazon highlights its cost-efficiency, claiming it’s around 80% more affordable than GPT-4o. Developers can leverage Nova Sonic via a new bidirectional streaming API on Bedrock for applications like customer service automation, voice assistants, and interactive learning.

IBM Launches z17 Mainframe: Powering the Enterprise AI Era

IBM has announced the IBM z17, its next-generation mainframe system, meticulously engineered for the age of artificial intelligence. Available from June 18th, 2025, the z17 integrates AI capabilities across hardware, software, and operations, powered by the new IBM Telum® II processor. This processor features increased frequency, compute capacity, and a 40% cache growth, enabling the z17 to perform 50% more AI inference operations per day than its predecessor, the z16, reaching over 450 billion inferencing operations daily with millisecond response times.

Designed with input from over 100 clients, the z17 supports over 250 AI use cases, from loan risk mitigation to medical image analysis. It enhances user experience by supporting AI assistants like IBM watsonx Code Assistant for Z and watsonx Assistant for Z. Furthermore, the upcoming IBM Spyre™ Accelerator (expected Q4 2025) will allow native running of IBM Granite-based models directly on the z17. Complementing the hardware, the z/OS 3.2 operating system (expected Q3 2025) and the new IBM Z Operations Unite offer (May 30th, 2025) will provide enhanced AI support and operational insights.

NVIDIA’s Llama 3.1 Nemotron Ultra 253B Surpasses DeepSeek R1 in Efficiency

NVIDIA has unveiled the Llama 3.1 Nemotron Ultra 253B, a large language model (LLM) that delivers exceptional performance despite its relatively compact size. With 253 billion parameters, it achieves results comparable to the 671 billion-parameter DeepSeek R1, outperforming it in tasks such as GPQA (76.01 vs. 71.5), IFEval instruction following (89.45 vs. 83.3), and LiveCodeBench coding tasks (66.31 vs. 65.9). This model demonstrates that efficient design can enable smaller models to rival larger counterparts, offering a favorable balance between accuracy and efficiency.

Nemotron Ultra leverages Neural Architecture Search (NAS) technology to reduce memory usage and inference latency significantly. It operates efficiently on an 8x NVIDIA H100 GPU node while supporting advanced reasoning tasks, human-interactive chat, and retrieval-augmented generation (RAG). The model also boasts a context length of up to 128K tokens, making it ideal for complex applications.

NVIDIA’s open-source approach via Hugging Face has sparked interest across the AI community, positioning Nemotron Ultra as a practical alternative to resource-intensive models like Llama4 Behemoth.

Google Enhances AI Mode with Complex Image Understanding

Google has significantly upgraded its experimental AI Mode in Search, enabling it to analyze images and answer complex, nuanced questions about their content. Leveraging Google Lens technology and a custom version of its Gemini AI model, this multimodal capability allows users to upload or capture a photo and ask detailed questions. The AI can understand the entire scene, including object relationships, materials, colors, shapes, and arrangements.

Using a “fan-out technique,” the AI Mode sends multiple queries about the image and its objects, generating comprehensive responses with links for deeper exploration. For example, users can ask for book recommendations similar to those pictured on a shelf. This feature aims to provide richer, more contextual information than standard searches, potentially requiring multiple queries otherwise. Initially available to Google One AI Premium subscribers in the US, access is expanding via the Google app (iOS/Android) for users enrolled in Labs, Google’s experimental platform.

Moonvalley: AI Video Startup Secures $43M in Funding

AI Video Startup Moonvalley Raises $43M Series B Moonvalley, a promising startup specializing in AI-powered video editing, has successfully closed a $43 million Series B funding round. This significant investment underscores the growing market for AI tools that streamline video production. Led by Insight Partners, the round also included participation from existing investors like Accel, Initialized Capital, and Quiet Capital.

Moonvalley’s platform leverages advanced AI algorithms to automate complex video editing tasks, such as automatic scene detection, shot selection, color grading, and even basic motion graphics. This technology empowers creators, from individual filmmakers to marketing professionals, to produce high-quality videos efficiently and cost-effectively. With this new funding, Moonvalley plans to expand its product offerings, enhance its AI capabilities, and further solidify its position in the burgeoning market for AI-driven video creation tools.

Krea Secures $83 Million to Revolutionize Generative AI for Creatives

San Francisco-based startup Krea has raised $83 million in funding to develop a unified platform that streamlines the use of multiple generative AI models for designers and visual creatives. The latest Series B round contributed $47 million, following earlier rounds totaling $36 million. This brings Krea’s post-money valuation to $500 million. The funding was led by Bain Capital Ventures, with participation from Andreessen Horowitz and Abstract Ventures.

Founded by Victor Perez and Diego Rodriguez, who met during their engineering studies in Barcelona, Krea aims to simplify the creative process by integrating various AI tools into a single, user-friendly interface. Their platform has attracted users from notable companies such as Perplexity AI, Loop Earplugs, Pixar, LEGO, and Samsung. The founders, both with backgrounds in AI research, notably declined postgraduate fellowships from the King of Spain to pursue their entrepreneurial vision.

IBM Acquires Hakkoda to Enhance AI Consulting Capabilities

In a strategic move to bolster its AI consulting services, IBM has acquired Hakkoda, a New York-based data and AI consultancy. This acquisition aims to expand IBM’s expertise in industries such as financial services, public sector, and healthcare and life sciences. Hakkoda specializes in assisting clients with migrating data to the cloud, particularly leveraging the Snowflake data cloud. The financial terms of the deal were not disclosed.

Founded in 2021 by former Deloitte General Manager Erik Duffield, Hakkoda had raised $5.6 million in venture capital from investors including Tercera, Lead Edge Capital, and Casimir Holdings. The company’s team of consultants across the U.S., Latin America, India, Europe, and the U.K. will integrate into IBM’s consulting division. This acquisition aligns with IBM’s ongoing investment in AI and automation technologies, following its recent purchases of DataStax and HashiCorp.

Microsoft Copilot Now Performs Web-Based Tasks and Personalized Actions

In celebration of its 50th anniversary, Microsoft has introduced powerful new features for its AI-driven chatbot Copilot, enabling it to browse websites and perform actions such as booking tickets or making reservations directly online. Copilot now remembers personalized details about users—similar to OpenAI’s ChatGPT—and can analyze live video from mobile devices to respond contextually based on observations.

These updates mark a significant step forward for Copilot, which historically lagged behind competitors like Google Gemini and ChatGPT in feature development. Microsoft has partnered with platforms such as Expedia, Kayak, OpenTable, TripAdvisor, and Skyscanner to ensure seamless integration across popular services. Additionally, Copilot can monitor online discounts for users and alert them when price changes occur.

Microsoft is reportedly redesigning Copilot with proprietary technology while maintaining compatibility with OpenAI models. These enhancements aim to position Copilot as a versatile assistant capable of simplifying everyday tasks through intuitive commands like “send flowers” or “track sales.” While promising, the effectiveness of these new capabilities remains under scrutiny as Microsoft continues refining its approach in response to user feedback.

Meta’s Llama 4 Launch Sparks Performance and Benchmark Debate

Around April 5th, 2025, Meta unveiled its latest family of open-weight large language models (LLMs), dubbed Llama 4. This release included multiple models: Llama 4 Scout (109B parameters, 16 experts, 17B active), Llama 4 Maverick (400B parameters, 128 experts, 17B active), and the still-in-training Llama 4 Behemoth (2T parameters, 16 experts, 288B active). Meta claimed significant performance, with Maverick reportedly outperforming models like GPT-4o on some benchmarks and Scout achieving a 10-million-token context window. These models are natively multimodal, processing text, images, and video, and utilize a mixture-of-experts (MoE) architecture.

Shortly after the launch, by April 6th, controversy arose regarding Llama 4 Maverick’s benchmark results, particularly on the popular Chatbot Arena leaderboard where it achieved a high ELO score (1417). Critics noted the benchmarked version was an “experimental chat version” (Llama-4-Maverick-03-26-Experimental) optimized for human preference, not the publicly released Hugging Face version. Allegations surfaced, including a disputed claim from an alleged employee, suggesting Meta might have “gamed the benchmarks” or trained on test sets. Meta’s VP of GenAI, Ahmad Al-Dahle, denied training on test sets, attributing performance variations to platform implementation needing stabilization. LMSYS, which runs Chatbot Arena, acknowledged Meta’s interpretation of policies didn’t match expectations and released battle data for transparency.

OpenAI Delays GPT-5 Release, Introduces Interim Models o3 and o4-mini

OpenAI CEO Sam Altman announced a change in plans via X (formerly Twitter), confirming the upcoming release of its anticipated reasoning models, o3 and o4-mini, likely within a few weeks. This reverses earlier indications that o3 might be integrated directly into the next flagship model. Consequently, the launch of GPT-5 has been postponed for a few months.

Altman cited several reasons for the delay, emphasizing the desire to make GPT-5 “much better than we originally thought”. He also acknowledged that integrating different components smoothly proved more challenging than expected and noted the need to ensure sufficient compute capacity to meet the “unprecedented demand” anticipated for GPT-5. This highlights the ongoing GPU constraints faced even by leading AI labs, which previously impacted GPT-4.5’s rollout and led to usage restrictions on features like image generation.

Okay, I will access the provided links and search the web for additional context to create the AI news summaries for the period of April 2nd to April 8th, 2025, following your outline and guidelines. I couldn’t access the specific webpages you linked. This sometimes happens due to paywalls, login requirements, or other access restrictions.

However, I was able to search the web for information on the topics mentioned in the URLs for the period around April 2nd to April 8th, 2025 (AI Week #14). Here’s a summary based on the information found, following your requested outline:

Meta’s Llama 4 Launch Sparks Performance and Benchmark Debate

Around April 5th, 2025, Meta unveiled its latest family of open-weight large language models (LLMs), dubbed Llama 4. This release included multiple models: Llama 4 Scout (109B parameters, 16 experts, 17B active), Llama 4 Maverick (400B parameters, 128 experts, 17B active), and the still-in-training Llama 4 Behemoth (2T parameters, 16 experts, 288B active). Meta claimed significant performance, with Maverick reportedly outperforming models like GPT-4o on some benchmarks and Scout achieving a 10-million-token context window. These models are natively multimodal, processing text, images, and video, and utilize a mixture-of-experts (MoE) architecture.

Shortly after the launch, by April 6th, controversy arose regarding Llama 4 Maverick’s benchmark results, particularly on the popular Chatbot Arena leaderboard where it achieved a high ELO score (1417). Critics noted the benchmarked version was an “experimental chat version” (Llama-4-Maverick-03-26-Experimental) optimized for human preference, not the publicly released Hugging Face version. Allegations surfaced, including a disputed claim from an alleged employee, suggesting Meta might have “gamed the benchmarks” or trained on test sets. Meta’s VP of GenAI, Ahmad Al-Dahle, denied training on test sets, attributing performance variations to platform implementation needing stabilization. LMSYS, which runs Chatbot Arena, acknowledged Meta’s interpretation of policies didn’t match expectations and released battle data for transparency.

OpenAI Confirms O3 Release, Pushes Back GPT-5 Launch

OpenAI CEO Sam Altman announced a change in plans via X (formerly Twitter) around April 4th, 2025, confirming the upcoming release of its anticipated reasoning models, o3 and o4-mini, likely within a few weeks. This reverses earlier indications that o3 might be integrated directly into the next flagship model. Consequently, the launch of GPT-5 has been postponed for a few months.

Altman cited several reasons for the delay, emphasizing the desire to make GPT-5 “much better than we originally thought”. He also acknowledged that integrating different components smoothly proved more challenging than expected and noted the need to ensure sufficient compute capacity to meet the “unprecedented demand” anticipated for GPT-5. This highlights the ongoing GPU constraints faced even by leading AI labs, which previously impacted GPT-4.5’s rollout and led to usage restrictions on features like image generation.

Midjourney Unveils V7: First Major Image Model Update in Over a Year

Midjourney, the independent research lab known for its AI image generator, announced the alpha release of its Midjourney V7 model around April 3rd, 2025. This marks the first major version update since V6 debuted over a year prior (December 2023). V7 introduces significant upgrades aimed at enhancing user experience and image quality, including default model personalization based on user visual preferences and new operational modes like “Draft Mode” for faster, cheaper rendering alongside existing Turbo and Relax modes.

While still in the alpha stage, V7 promises improvements over V6, with the lab planning frequent updates over the next 60 days based on community feedback gathered via Discord. Future V7 updates are expected to enhance editing, upscaling, and retexturing capabilities, though these currently rely on V6. Midjourney noted that V7 might require different prompting styles compared to previous versions. A new “Voice Mode” feature with text-to-talk capabilities was also mentioned as an enhancement.

OpenAI Makes Inaugural Cybersecurity Investment in Adaptive Security

Marking its first foray into funding the cybersecurity sector, the OpenAI Startup Fund, alongside renowned venture capital firm Andreessen Horowitz (a16z), participated in an investment round for Adaptive Security. Reports around April 3rd-4th, 2025 indicated the total investment was $43 million, though the exact breakdown between investors wasn’t specified in the search results.

Adaptive Security employs a unique strategy, utilizing AI to simulate sophisticated cyberattacks, including AI-generated phishing attempts via calls or emails. This allows the platform to train users and systems to better detect and respond to AI-driven threats in real-time. Brian Long, CEO and co-founder of Adaptive Security, stated the funding will be used for hiring and further system development. This investment aligns with OpenAI’s broader efforts to address AI safety and security risks, including forming internal teams and partnering with defense tech companies.

Intel and TSMC Reportedly Form Joint Venture to Revitalize U.S. Chip Manufacturing

Intel and Taiwan Semiconductor Manufacturing Company (TSMC) have tentatively agreed to establish a joint venture aimed at operating Intel’s semiconductor manufacturing facilities in the United States. According to reports, TSMC is set to acquire a 20% stake in this new entity. This collaboration emerges as Intel seeks to overcome recent challenges, including an $18.8 billion net loss in 2024, marking its first annual net loss since 1986. The partnership is anticipated to infuse much-needed expertise and capital into Intel’s operations, aligning with U.S. government efforts to bolster domestic semiconductor production.  

Market reactions to the announcement have been mixed. Intel’s stock experienced an 8.2% decline, closing at $20.60, influenced by broader market concerns such as China’s imposition of 34% tariffs on U.S. goods. Despite these immediate market fluctuations, analysts suggest that TSMC’s involvement could provide Intel with critical technological advancements and operational efficiencies, potentially enhancing its competitiveness in the semiconductor industry. ​

Phonic Secures $4 Million Seed Funding from Lux Capital to Advance Voice AI Solutions

Phonic, an AI voice technology startup, has raised $4 million in a seed funding round led by Lux Capital, with participation from notable investors including Amjad Masad (co-founder of Replit), Clem Delangue (co-founder of Hugging Face), Qasar Younis (co-founder of Applied Intuition), and Erik Bernhardsson (founder of Modal Labs). Founded by MIT graduates Moin Nadeem and Nikhil Murthy, Phonic aims to deliver an end-to-end voice AI stack that enhances the reliability and reduces the latency of synthetic voice applications.

Phonic differentiates itself by developing proprietary models trained in-house, allowing for deep integration of reliability features and cost-efficient operations. The company’s approach involves training models on diverse recordings, including those with accented and muffled speech, to ensure robustness. Currently collaborating with partners in the insurance and healthcare sectors, Phonic plans to broaden its product availability in the coming months, aiming to set new standards in the voice AI industry.

Runway Raises $308 Million to Expand AI-Powered Video Generation Platform

Runway, renowned for its AI-driven video generation models, has secured $308 million in a Series D funding round led by General Atlantic, with contributions from Fidelity Management & Research Company, Baillie Gifford, Nvidia, and SoftBank. This investment elevates the New York-based company’s total funding to $536.5 million, underscoring its significant impact in the AI media production sector.

The newly acquired capital will be directed towards advancing AI research, expanding the team, and enhancing Runway’s film and animation production division, Runway Studios. The company’s latest model, Gen-4, boasts capabilities such as creating consistent characters and environments across scenes, maintaining coherent world settings, and regenerating elements from various perspectives. Despite facing competition from industry giants like OpenAI and Google, Runway’s strategic partnerships, including collaborations with major Hollywood studios, position it as a formidable player in the AI media landscape.

Actively AI Raises $22.5 Million to Revolutionize Sales with ‘Superintelligence’

Actively AI, a startup specializing in AI-driven sales solutions, has announced a $22.5 million funding round, comprising a $17.5 million Series A led by Bain Capital Ventures and a previously undisclosed $5 million seed round from First Round Capital. Founded by Stanford alumni Anshul Gupta and Mihir Garimella, Actively AI focuses on developing custom “reasoning” models that analyze company data to identify high-value sales prospects, aiming to emulate the decision-making processes of top human sales representatives.

Actively AI’s approach contrasts with traditional AI sales tools that prioritize volume over quality. By leveraging a combination of proprietary models and technologies from OpenAI and Anthropic, the company seeks to provide a more nuanced and effective sales strategy. Early collaborations with clients like Ramp have reportedly resulted in substantial revenue increases, highlighting the potential of Actively AI’s innovative methodology in transforming sales operations. As we conclude this week’s exploration of AI’s transformative developments, it’s evident that the synergy between technological innovation and strategic investments continues to drive the industry forward. Stay tuned for next week’s insights as we delve into the latest breakthroughs and trends shaping the future of artificial intelligence.

Leave a Comment

Your email address will not be published. Required fields are marked *