Why Try AI

Why Try AI

Share this post

Why Try AI
Why Try AI
Sunday Rundown #95: So Much Google & So Much Waldo
Copy link
Facebook
Email
Notes
More
Sunday Rundown

Sunday Rundown #95: So Much Google & So Much Waldo

Sunday Bonus #55: Comparison page for 15 AI video models.

Daniel Nest's avatar
Daniel Nest
Apr 13, 2025
∙ Paid
14

Share this post

Why Try AI
Why Try AI
Sunday Rundown #95: So Much Google & So Much Waldo
Copy link
Facebook
Email
Notes
More
6
2
Share

Heads up: I’m traveling for Easter with my family, so the next Sunday Rundown will be on April 27.

Happy Sunday, friends!

Welcome back to the weekly look at generative AI that covers the following:

  • Sunday Rundown (free): this week’s AI news + a fun AI fail.

  • Sunday Bonus (paid): an exclusive segment for my paid subscribers.

Every Sunday Bonus in one place

Let’s get to it.

🗞️ AI news

Here are this week’s AI developments.

👩‍💻 AI releases

New stuff you can try right now:

  1. Amazon news (models available via Amazon Bedrock):

    1. Nova Sonic is a speech model that can pick up subtle cues like tone, pacing, inflection, and more.

    2. Nova Reel 1.1 is a video model that can create 2-minute multi-shot clips from text and image prompts.

  2. Anthropic introduced a two-tiered Max Plan for Claude, with 5x more usage ($100/month) or 20x more usage ($200/month) plus early access to new features.

  3. Canva rolled out Visual Suite 2.0, a massive product upgrade with dozens of new features, many powered by AI.

  4. Deep Cogito released Cogito v1 Preview, a family of open-source LLMs that outperform competitor models of comparable sizes.

  5. Google news (will Google ever take a break?):

    1. AI Mode is now multimodal, so you can snap a photo or upload an image to ask questions and get comprehensive responses.

    2. Deep Research is now powered by Gemini 2.5 Pro Experimental for Gemini Advanced users, making it a far more capable research agent that now convincingly outperforms OpenAI’s version.

    3. Firebase Studio uses Gemini models to help developers prototype, build, and deploy full-stack applications more efficiently. (Try it here.)

    4. Gemini Live now lets users share their phone’s screen and camera during conversations for rich, multimodal chats. (Previously known as “Project Astra”)

    5. Vertex AI now incorporates all of Google’s creative AI models like Lyria (music), Veo 2 (video), Imagen 3 (images), and Chirp 3 (voice), most of which have also received additional improvements and features.

    6. Workspace product suite is getting many new AI capabilities, tied together by the new Workspace Flows, which helps orchestrate work across different apps.

    7. …and even more stuff announced at Google Cloud Next 2025.

  6. Meta released Llama 4, the next generation of its open-source, natively multimodal family of language models. (Reception has been mixed.

    Charlie Guo
    wrote an excellent summary.)

  7. Microsoft rolled out almost a dozen new Copilot features including memory, vision, agentic tasks, and more.

  8. Moonshot AI open-sourced Kimi-VL and Kimi-VL-Thinking, lightweight vision-language models that excel in multimodal reasoning.

  9. NVIDIA’s new Llama 3.1 Nemotron Ultra 253B excels at advanced reasoning and instruction following, suited for high-accuracy scientific, code, and math tasks.

  10. OpenAI expanded ChatGPT’s Memory to (optionally) reference all past chats to give personalized responses. Rolling out to Pro and Plus accounts outside the EU.

  11. Runway introduced Gen-4 Turbo, a faster version of Gen-4 Alpha that can generate 10-second clips in just 30 seconds. (Also available on free accounts, as long as you have credits left.)

  12. WordPress launched an AI Website Builder that designs complete sites with images and text from a simple prompt.

  13. YouTube is rolling out a Music Assistant that generates music tracks from prompts to a subset of Creator Music users.


🔬 AI research

Cool stuff you might get to try one day:

  1. Adobe is working on a range of AI agents that can intelligently assist users of its products: Acrobat, Express, Photoshop, and Premiere Pro.

  2. Researchers NVIDIA and several US universities outlined a video generation method that uses Test-Time Training to create coherent 1-minute video stories.


📝 Suddenly, a surprise survey spawns…

Please help make Why Try AI better. Let me know what works and what doesn’t:

Share your feedback


🤦‍♂️ AI fail of the week

Waldo is somewhere in this GPT-4o image. See if you can find him.

A "Where's Waldo" style iamge with dozens of Waldos

Why Try AI is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.


💰 Sunday Bonus #55: Interactive page to compare 15 AI video models

Wow, last week’s GPT-4o use cases swipe file proved quite popular!

So I once again went to my new friend Genspark Super Agent to help turn my research insights into a neat interactive page. This time, it’s about video models.

While working on my recent guest post about AI video for

AI Supremacy
, I put together a Google Sheet comparing different video models, pricing, and features.

This week, I spent a bit of time with Genspark and Gemini to turn my initial research into a nice page that lets you explore and compare these models:

AI Video Models Explorer

If nothing else, it’ll give you a great overview of the current AI video landscape and available models.

This post is for paid subscribers

Already a paid subscriber? Sign in
© 2025 Daniel Nest
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More