top of page
Another Crazy Day in AI: An Almost Daily Newsletter

Hello, AI Enthusiasts.


Just checking in after the day’s grind. Here’s a quick drop.


Google’s Gemini 2.5 is pushing AI audio chat to new heights—making human-machine talk smoother than ever. The machines are getting chatty, aren’t they?


Meanwhile, OpenAI’s going big with a billion-dollar bet on hardware. So yep, expect new gadgets that rethink how we tech.


But it’s not all smooth sailing—FDA’s AI tools are rough around the edges. Guess humans aren’t out of a job just yet.


Here's another crazy day in AI:

  • Google DeepMind's audio tech gets more intuitive

  • OpenAI bets big on physical AI

  • FDA AI rollout stumbles

  • Some AI tools to try out


TODAY'S FEATURED ITEM: Gemini Learns to Read Vocal Cues


A robotic scientist in a classic white coat with 'AI Scientist' on its back stands beside a human scientist with 'Human Scientist' on their coat, looking towards the AI Scientist.

Image Credit: Wowza (created with Ideogram)


What if machines could not only understand what you say, but how you say it?


Google researchers Ankur Bapna and Tara Sainath recently shared insights into how Gemini 2.5 approaches audio in ways that go beyond basic speech recognition. Their work, published through Google DeepMind, focuses on native audio capabilities that interpret more than just words—they pay attention to the subtle elements of human communication, such as tone, pace, emotional undertones, and even the pauses we use strategically in conversation. These advancements are already being integrated into tools like NotebookLM’s Audio Overviews and Project Astra, signaling a new way machines handle spoken language.



Here are some of the notable aspects Gemini 2.5 brings to audio processing:

  • Processes conversations in near real-time to keep interactions fluid

  • Filters out background noise to focus on the speaker’s voice

  • Understands more than 24 languages and can handle mixed-language speech

  • Recognizes emotional signals such as tone and emphasis to better grasp meaning

  • Offers customizable speech synthesis including style, emotion, accent, and speed

  • Generates multi-person dialogues from written scripts

  • Combines audio and visual cues for richer context understanding

  • Includes audio watermarking (SynthID) to identify AI-generated speech

  • Provides developers with tools to fine-tune speech and dialogue control



The practical implications of these developments extend beyond convenience features. People who process information better through audio or face challenges with traditional text interfaces may find these advances particularly valuable. Content creators could discover new efficiencies in producing audio materials, while educators might develop more engaging ways to present complex information through conversational formats. The technology also opens possibilities for more accessible customer service interactions and personalized learning experiences.



However, these same capabilities raise important questions about authenticity and disclosure in digital communication. As synthetic voices become more emotionally nuanced and contextually aware, distinguishing between human and machine-generated content becomes increasingly complex. While the inclusion of watermarking technology demonstrates awareness of these concerns, the practical challenges of maintaining transparency in everyday interactions remain significant. The development represents meaningful progress in making technology more intuitive and responsive, yet it also underscores the need for thoughtful approaches to implementation and user awareness. As voice technology becomes more sophisticated, balancing innovation with ethical considerations will likely require ongoing attention from both developers and users.




Read the full blog here.

OTHER INTERESTING AI HIGHLIGHTS:


OpenAI Bets Big On Physical AI

/John K. Waters, Editor in Chief, on Campus Technology


OpenAI is making its biggest acquisition yet: a $6.5 billion all-stock deal to acquire io, an AI hardware startup co-founded by former Apple design chief Jony Ive. The move marks a major push into AI-powered consumer devices, with the goal of building products that rethink how we interact with technology—beyond the smartphone or laptop. With design led by Ive’s LoveFrom studio and development by OpenAI’s product team, the first devices are expected to launch in 2026. Industry analysts see this as OpenAI’s official entry into “physical AI.”



Read more here.


FDA AI Rollout Stumbles

/Berkeley Lovelace Jr., Health and Medical Reporter, on NBC News


The FDA is pushing forward with AI-powered tools to help streamline regulatory tasks, but internal sources say the technology isn't ready. A medical device review assistant known as CDRH-GPT is reportedly buggy, disconnected from live databases, and unable to answer basic questions. A second tool, Elsa, is now being used agency-wide, but staff report inaccurate summaries and growing concerns that the rollout is too rushed. Experts warn that AI’s use in high-stakes regulatory decisions must be carefully balanced with human oversight.



Read more here.

SOME AI TOOLS TO TRY OUT:


  • Writegenic – Instantly generate project and business documents with AI.

  • Chronicle – Create designer-grade presentations effortlessly with AI.

  • Manus Slides – Build complete slide decks from a single prompt.


That’s a wrap on today’s Almost Daily craziness.


Catch us almost every day—almost! 😉

EXCITING NEWS:

The Another Crazy Day in AI newsletter is on LinkedIn!!!



Wowza, Inc.

Leveraging AI for Enhanced Content: As part of our commitment to exploring new technologies, we used AI to help curate and refine our newsletters. This enriches our content and keeps us at the forefront of digital innovation, ensuring you stay informed with the latest trends and developments.





Another Crazy Day in AI: An Almost Daily Newsletter

Hello, AI Enthusiasts.


Made it through the first workday? Same. Meanwhile, AI’s been busy.


Microsoft just dropped a mobile version of its Bing Video Creator—powered by Sora—letting anyone with a Microsoft account whip up short, AI-generated videos right from their phone.


Meta’s pushing the envelope too, working on AI that could handle Facebook and Instagram ads start to finish. Imagine telling AI your goals and budget—and letting it handle copy, visuals, targeting, and placement.


Also: Chinese researchers used AI to pinpoint a way to revive dead EV batteries—potentially slashing waste and cost.


Here's another crazy day in AI:

  • Sora now available for free through Bing Video Creator mobile

  • Meta pushes AI deeper into advertising

  • AI discovers fix for dead EV batteries

  • Some AI tools to try out


TODAY'S FEATURED ITEM: Sora in Your Pocket


A robotic scientist in a classic white coat with 'AI Scientist' on its back stands beside a human scientist with 'Human Scientist' on their coat, looking towards the AI Scientist.

Image Credit: Wowza (created with Ideogram)


Have you ever wondered when AI video generation would become truly accessible to everyone, not just those willing to pay premium subscription fees?


Microsoft recently launched a mobile version of its Bing Video Creator, quietly introducing a way for anyone with a Microsoft account to experiment with short, AI-generated videos. The tool is powered by Sora and available on both iOS and Android.


In an article published by VentureBeat, Executive Editor Carl Franzen takes a closer look at how this rollout adds a new entry point for people curious about AI video generation. While the outputs are short, just five seconds and limited to vertical format, they’re freely available, quick to create, and don’t require any technical experience.



A few things to know about the experience:

  • Access is through the Bing app (no separate download needed)

  • Generates 5-second vertical videos from text prompts

  • First 10 generations are prioritized for speed

  • Unlimited slower generations available after the first 10

  • Horizontal format isn’t available yet, but it’s in the works

  • Videos can be downloaded, shared, or linked directly

  • No subscription required — just a Microsoft account

  • Extra fast generations can be earned through Microsoft Rewards

  • Safety tools and moderation are integrated into the platform

  • The feature is positioned more as an experiment than a polished product


The experience is minimal by design. There are no advanced controls, editing tools, or cinematic flourishes. Instead, this is a chance to observe a baseline: where generative video tools currently stand when placed in the hands of everyday users. Five-second clips aren’t going to replace traditional media. That’s not the point. But they do offer a glimpse into how accessible video creation might evolve from here.


For educators, hobbyists, or anyone casually curious about generative media, the barrier to entry here is exceptionally low. There’s nothing to buy, no software to install beyond the Bing app, and no real risk in exploring what a few words can become. And perhaps more interestingly, this rollout wasn’t surrounded by marketing fanfare. That quietness might suggest Microsoft is observing more than promoting, seeing how people actually use the tool, what prompts they try, and whether these small moments of creativity hint at something bigger.


The tool may not feel essential, at least not yet, but it’s worth noting where it lives: quietly embedded in a search app many already have. That subtle integration could say more about where everyday AI is heading than the tool itself.




Read the full article here.

OTHER INTERESTING AI HIGHLIGHTS:


Meta Pushes AI Deeper Into Advertising

/Macy Meyer, Writer II, on CNET


Meta is reportedly working on AI systems capable of fully automating the creation and management of ads on Facebook and Instagram. This initiative would let businesses simply share their objectives and budget, while AI generates copy, visuals, targeting, and media placement—potentially without human input. Meta says this will benefit small businesses most, by reducing time and costs, though critics are raising concerns over misinformation, bias, and job impacts in marketing and creative industries. With AI already infused across Meta’s platforms, this latest step signals a larger push to reshape advertising as we know it.



Read more here.


AI Discovers Fix For Dead EV Batteries

/You Xiaoying, Freelance Journalist, edited by Andrea Thompson (Associate Editor), on Scientific American


Chinese researchers have developed a machine learning-guided method that revives dead lithium-ion EV batteries using a specialized chemical injection, restoring up to 96% capacity even after 12,000 charge cycles. The AI model identified lithium trifluoromethanesulfinate as a promising molecule to “recharge” degraded battery cells, a breakthrough that could dramatically reduce battery waste and production needs. While tests are currently limited to individual cells, efforts are underway to scale the process to full battery packs. Experts say the concept could be commercially viable but will face engineering and safety challenges.



Read more here.

Lithium ions were restored to a spent battery pack by injecting an electrolyte solution. Zhao Chihao
Lithium ions were restored to a spent battery pack by injecting an electrolyte solution. Zhao Chihao

SOME AI TOOLS TO TRY OUT:


  • Duckbill – Handles bookings, bills, and errands using AI + human help.

  • Zero – AI-native email client that manages your inbox for you.

  • OpenPaper – Chat with research papers and get cited answers from the source.


That’s a wrap on today’s Almost Daily craziness.


Catch us almost every day—almost! 😉

EXCITING NEWS:

The Another Crazy Day in AI newsletter is on LinkedIn!!!



Wowza, Inc.

Leveraging AI for Enhanced Content: As part of our commitment to exploring new technologies, we used AI to help curate and refine our newsletters. This enriches our content and keeps us at the forefront of digital innovation, ensuring you stay informed with the latest trends and developments.





Another Crazy Day in AI: An Almost Daily Newsletter

Hello, AI Enthusiasts.


If you’ve been hopping between meetings or running on low battery—this update might just recharge your brain.


Video calls just got a new twist—imagine an AI avatar that shows your face and expressions without you having to be on camera. Real-time, expressive, and even multilingual. The future of “showing up” might not require you at all.


Meanwhile, a new podcast series takes a serious look at AI’s impact in schools and society. And in the workplace? AI’s raising output, but will paychecks follow? The answer might surprise you.


Take these brain snacks with you into the weekend.


Here's another crazy day in AI:

  • The rise of digital presence

  • AI Education Project opens studio for AI learning talks

  • Will AI widen the productivity-pay gap

  • Some AI tools to try out


TODAY'S FEATURED ITEM: Video Without the Webcam


A robotic scientist in a classic white coat with 'AI Scientist' on its back stands beside a human scientist with 'Human Scientist' on their coat, looking towards the AI Scientist.

Image Credit: Wowza (created with Ideogram)


In what kinds of situations would a live avatar be helpful—and when might it feel out of place?


In a time when video calls are an everyday part of work, learning, and social connection, the idea of presenting yourself digitally is getting a rethink. In a recent USA Today piece, contributor Matthew Kayser introduces us to Akool Live Camera—a tool that uses artificial intelligence to generate real-time, realistic avatars of people during live video interactions. Whether you're in a Zoom meeting, hosting a class, or collaborating across time zones, the technology offers an alternative to traditional webcams, raising important questions about presence, identity, and communication.


The Live Camera tool creates an avatar that mirrors your expressions and voice in real time, and can also translate speech into other languages mid-conversation. It’s designed to work across major platforms and is being positioned as useful for anyone who might prefer not to appear on camera directly—whether for personal comfort, privacy, or accessibility. But alongside these functions, it invites broader questions about how digital communication tools are evolving, and how much realism or “liveness” we expect when we interact through a screen.



Some key features to be aware of

  • Works in real time, creating a photorealistic digital version of you that mirrors your facial movements

  • Compatible with common video platforms like Zoom, Google Meet, and Microsoft Teams

  • Includes near-instant voice translation and lip-sync in multiple languages

  • Offers voice cloning for continuity and expressiveness

  • Can be used in live meetings, classrooms, interviews, or even content creation

  • Free during its beta period, with multilingual and privacy controls in place

  • Created by Dr. Jiajun (Jeff) Lu and the Akool team, focused on digital communication tools



What makes this development notable is less about the technology itself and more about how it’s being introduced into familiar, everyday spaces. Video communication has long relied on cameras as a default, but this tool opens up another route, one that’s not anonymous, but not entirely “live” either. Depending on the context, that in-between space could be helpful or even necessary. In others, it might feel out of sync with the tone of the conversation or the expectations of the audience.


For many users, especially those navigating language barriers or feeling uneasy about being on camera, a real-time avatar could make it easier to participate fully. At the same time, it raises questions around transparency and how we perceive authenticity in digital interactions. If the person speaking looks slightly different or isn’t live on screen in the usual sense, does that change how we listen or how we respond? These aren’t questions with one answer, but they’re worth keeping in mind as tools like these continue to evolve and find new use cases in work, education, and beyond.




Read the full article here.

Check it out here.

OTHER INTERESTING AI HIGHLIGHTS:


AI Education Project Opens Studio For AI Learning Talks

/News Staff, on Government Technology


The AI Education Project (aiEDU), a nonprofit focused on building AI readiness in education, has launched aiEDU Studios, a new video podcast platform featuring in-depth weekly conversations with experts on AI’s evolving role in schools, work, and society. Hosted by aiEDU co-founder and CEO Alex Kotran, the long-form episodes aim to move beyond soundbites and into deeper discussions about AI implementation. Topics range from data literacy to classroom collaboration and include insights from thought leaders like Kumar Garg and Khushali Narechania. The podcast is part of aiEDU’s broader mission to prepare students and educators for a future shaped by AI.



Read more here.


Will AI Widen The Productivity-Pay Gap?

/John Letzing (Digital Editor, Economics), on World Economic Forum


As AI continues to transform the workplace, questions about whether rising productivity will translate into rising pay remain unresolved. In a new article for the World Economic Forum, John Letzing explores the widening productivity-pay gap and the role AI could play in either bridging or worsening it. While AI promises to boost efficiency—particularly for newer or lower-paid workers—it also threatens to reduce the wage premium for experienced knowledge workers. Chief economists warn that without proactive policies like upskilling and worker protections, AI may deepen inequality even as it improves output.



Read more here.

Chief economists weighed in on AI's anticipated impacts. Image Credit: World Economic Forum
Chief economists weighed in on AI's anticipated impacts. Image Credit: World Economic Forum

SOME AI TOOLS TO TRY OUT:


  • Wondera – Write, produce, and publish songs just by chatting with AI.

  • Parabola – Build AI workflows to clean and organize data from PDFs, emails, and more.

  • Tyce – A modern, minimalist alternative to Google Docs built for focused writing.


That’s a wrap on today’s Almost Daily craziness.


Catch us almost every day—almost! 😉

EXCITING NEWS:

The Another Crazy Day in AI newsletter is on LinkedIn!!!



Wowza, Inc.

Leveraging AI for Enhanced Content: As part of our commitment to exploring new technologies, we used AI to help curate and refine our newsletters. This enriches our content and keeps us at the forefront of digital innovation, ensuring you stay informed with the latest trends and developments.





Copyright Wowza, inc 2025
bottom of page