🤖 Meta just bought Manus AI

PLUS: How to turn a static design into organic, cinematic motion for free

Good Morning! Meta has just acquired Manus, a startup known for building general-purpose AI agents. Plus, I'll show you how to turn a static design into organic, cinematic motion for free

In today’s AI newsletter:

  • Meta Acquires Manus to Scale AI Agents

  • How to Turn a Static Design into Organic, Cinematic Motion

  • YouTube Is Rewarding AI Slop

  • Nvidia Licenses Groq’s Inference Tech (Without Acquiring It)

  • 4 new AI tools worth trying

AI PARTNERSHIPS

Meta has acquired Manus, a startup known for building general-purpose AI agents that can automate research, workflows, and multi-step tasks.

  • Manus agents handle end-to-end work, not just suggestions

  • Processed 147T+ tokens and powered 80M+ virtual computers

  • Used by both individuals and enterprises for scalable automation

  • Manus will continue operating from Singapore, with subscriptions still live

Manus gives Meta a strong “do the work for you” layer, helping it compete with AI-first productivity platforms. If integrated across Meta’s apps, these agents could reach billions of users and turn AI from an assistant into actual digital labor.

AI SLOP

A new study shows YouTube is actively pushing low-effort, AI-generated content to users, especially new ones.

  • 104 of the first 500 videos shown to new users were pure AI slop

  • 278 major channels post nothing else, pulling 63B views and 221M subs

  • Estimated $117M/year generated from this content alone

  • Videos are optimized for silent, late-night, endless scrolling

The algorithm rewards volume, repetition, and familiar weirdness over originality. That money flows fast, especially in regions where even small payouts beat local wages, while course sellers promise shortcuts. If 20% of a fresh feed is already slop, the real concern is what happens once YouTube learns exactly how much low quality you’ll accept.

AI PARTNERSHIPS

Nvidia is paying roughly $20B for non-exclusive access to Groq’s inference chip IP, while also hiring key Groq leaders, but Groq stays independent.

  • Nvidia gets Groq’s low-latency inference architecture without a full acquisition

  • Key Groq talent joins Nvidia to plug that know-how into its “AI factory” stac

  • Groq continues running its cloud business under new CEO Simon Edwards

  • Structure avoids heavy regulatory scrutiny tied to outright M&A

AI is shifting from training giant models to serving billions of real-time queries, where latency, cost per response, and memory bottlenecks matter more than raw compute. Groq is optimized for fast, predictable inference, and Nvidia is hedging early. This deal shows how Big Tech can absorb disruptive tech: license the IP, hire the brains, skip the regulatory blast radius.

HOW TO AI

🗂️ How to Turn a Static Design Into Organic, Cinematic Motion

In this tutorial, you’ll learn how to use an AI-powered motion canvas to transform a simple logo or static image into fluid, organic visuals using shaders, presets, and natural-feeling motion.

🧰 Who is This For

  • Creators who want studio-quality visuals fast

  • Marketers building product or lifestyle content

  • Designers who don’t want to prompt-engineer everything

  • Anyone who wants consistent results with minimal effort

STEP 1: Access the Tool

Head over to paper.design and sign up for an account. Once you’re logged in, click the “New Project” button on your dashboard to start fresh.

You’ll land directly inside the main workspace. On the left, you’ll see the Layers panel, which controls everything on your canvas. On the right is the Properties panel, where you fine-tune motion and effects. At the bottom of the screen, you’ll notice the AI Prompt bar, which is used to generate vibe-based visuals.

STEP 2: Upload and Position Your Design

To bring in your design, simply drag and drop your logo or static image directly onto the canvas.

Once the image is visible, click on its layer in the left-hand panel to make sure it’s selected. Use the corner handles on the canvas to resize, reposition, or center your design so it sits cleanly in the frame and is ready for motion effects.

STEP 3: Apply Shaders and Presets

Open the shaders section from the left area of the interface. Inside, you’ll find preset categories such as Organic, Liquid, and Glass. Clicking any preset instantly applies a shader-based effect to your design.

STEP 4: Generate a Vibe and Fine-Tune the Motion

At the bottom of the screen, click into the AI Prompt bar and describe the mood you want. You can write something like “deep sea bioluminescence” or “molten liquid gold” and press Enter. The AI generates visuals that match the emotional tone of your description.

To refine the result, move to the Properties panel on the right. Here you can adjust motion speed to control energy, intensity to strengthen the effect, and motion flow to guide the direction of movement.

SoftBank agrees to acquire NYSE-listed DigitalBridge, a PE firm that invests in data centers and digital infrastructure operators, for $4B, including debt

Z AI, a company behind GLM model family, is going for an IPO on Jan 8 and set to raise $560 million.

Notion is working on AI-first Workspaces to enable users to build AI-powered organisations. This feature is expected to arrive soon in Early Access.

Cursor CEO Michael Truell warns that “vibe coding” advanced projects may create “shaky foundations” and eventually “things start to kind of crumble”.

🎨GPT Image 1.5: OpenAI’s latest, faster and more controllable image model

🧠 Oboe: Instantly create AI-powered learning courses

🤖 Nemotron 3: Nvidia’s open-source models for building multi-agent systems

🤝 Lumro: Deploy AI agents for top-notch customer experience

Which image is real?

Login or Subscribe to participate in polls.

THAT’S IT FOR TODAY

Thanks for making it to the end! I put my heart into every email I send, I hope you are enjoying it. Let me know your thoughts so I can make the next one even better!

See you tomorrow :)

- Dr. Alvaro Cintas