Thursday, November 20, 2025
spot_img
More
    HomeLatest NewsGoogle Unveils Gemini 3, Model Capable of Analysing Uploaded Videos

    Google Unveils Gemini 3, Model Capable of Analysing Uploaded Videos

    Google has introduced Gemini 3, the latest version of its multimodal artificial intelligence model, describing it as a major step forward in the company’s development of general-purpose AI systems. The model was announced with upgrades in reasoning, context understanding and automated task execution across Google’s consumer and developer platforms.

    Also read: Is Google the New King of AI?

    According to the company, Gemini 3 features state-of-the-art reasoning capabilities, deeper multimodal processing across text, images and other inputs, and what Google refers to as “vibe coding,” aimed at enabling users to move from a single prompt to a functional application. The model also includes expanded agentic features that allow it to carry out multi-step tasks on behalf of users.

    Also read: Top 10 Trending Gemini AI Photo Prompt Commands

    Google said the model has been designed to better interpret context and intent, reducing the amount of prompting required. The company noted significant growth in usage across its AI ecosystem, stating: “AI Overviews now have 2 billion users every month. The Gemini app surpasses 650 million users per month, more than 70% of our Cloud customers use our AI, 13 million developers have built with our generative models, and that is just a snippet of the impact we’re seeing.”

    Also read: Can Google’s Med-Gemini AI Change the Face of Healthcare?

    Sundar Pichai, CEO of Google and Alphabet, said Gemini 3 represents a substantial evolution in the company’s AI systems. “It’s the best model in the world for multimodal understanding, and our most powerful agentic plus vibe coding model yet. Gemini 3 can bring any idea to life, quickly grasping context and intent so you can get what you need with less prompting. Find Gemini 3 Pro rolling out today in the Gemini app and AI Mode in Search. For developers, build with it now in Google AI Studio and Vertex AI. Excited for you to try it!” he said.

    Gemini 3 extends its multimodal capabilities by processing images, documents and sketches, allowing users to generate outputs ranging from websites to games. Google said the model can convert a rough drawing into a functional interface or transform diagrams into interactive lessons.

    The company also highlighted enhancements in multimodal reasoning, enabling the model to analyze long-form sports videos and produce detailed performance assessments, including identification of technique issues and recommended corrective drills.

    In Search, Gemini 3 supports new generative layouts that combine text, images and interactive elements. Users querying complex scientific concepts, such as the physics of the three-body problem, may receive visual simulations and structured explanations. Google said these layouts are intended to offer a magazine-like experience with customizable elements.

    As part of the rollout, Google introduced Gemini Agent, a feature that uses the model’s reasoning capabilities to divide complex tasks into actionable steps. The agent can suggest actions such as drafting email responses or archiving messages by leveraging integrations across Google applications. The feature will be available on the web for Google AI Ultra subscribers in the United States. Gemini 3 Pro begins rolling out today in the Gemini app and AI Mode in Search, with developer access available through Google AI Studio and Vertex AI.

    Author

    RELATED ARTICLES

    LEAVE A REPLY

    Please enter your comment!
    Please enter your name here

    Most Popular

    spot_img
    spot_img