Meta Tests Image-Generating Social Chatbot on Its Platforms

Meta is testing a new large language chatbot, Meta AI, on social platforms in parts of India and Africa. The chatbot was introduced in late 2023, and began testing on U.S. WhatApp users in March. The test is expanding to include more territories and the addition of Instagram and Facebook Messenger. India is reported to be Meta’s largest social market, with more than 500 million Facebook and WhatsApp users, and has big implications as the company scales up its AI plans to compete against OpenAI and others. The Meta AI chatbot answers questions and generates photorealistic images. Continue reading Meta Tests Image-Generating Social Chatbot on Its Platforms

Google Offers Public Preview of Gemini Pro for Cloud Clients

Google is moving its most powerful artificial intelligence model, Gemini 1.5 Pro, into public preview for developers and Google Cloud customers. Gemini 1.5 Pro includes what Google claims is a breakthrough in long context understanding, with the ability to run 1 million tokens of information “opening up new possibilities for enterprises to create, discover and build using AI.” Gemini’s multimodal capabilities allow it to process audio, video, text, code and more, which when combined with long context, “enables enterprises to do things that just weren’t possible with AI before,” according to Google. Continue reading Google Offers Public Preview of Gemini Pro for Cloud Clients

Opera Browser Is Experimenting with Local Support for LLMs

Opera has become the first browser to add support for large language models (LLMs). At this point the feature is experimental, and available only on the Opera One Developer browser as part of the AI Feature Drops program. The update offers about 150 LLMs from more than 50 different families, including Meta’s LLaMA, Google’s Gemma, Mixtral and Vicuna. Opera had previously only offered local support for its own Aria AI, a competitor to Microsoft Copilot and OpenAI’s ChatGPT. The local LLMs are being offered for testing as a complimentary addition to Opera’s online Aria service. Continue reading Opera Browser Is Experimenting with Local Support for LLMs

Apple’s ReALM AI Advances the Science of Digital Assistants

Apple has developed a large language model it says has advanced screen-reading and comprehension capabilities. ReALM (Reference Resolution as Language Modeling) is artificial intelligence that can see and read computer screens in context, according to Apple, which says it advances technology essential for a true AI assistant “that aims to allow a user to naturally communicate their requirements to an agent, or to have a conversation with it.” Apple claims that in a benchmark against GPT-3.5 and GPT-4, the smallest ReALM model performed “comparable” to GPT-4, with its “larger models substantially outperforming it.” Continue reading Apple’s ReALM AI Advances the Science of Digital Assistants

Microsoft, OpenAI Considering a Supercomputer Data Center

Microsoft and OpenAI are contemplating an AI supercomputer data center that may cost as much as $100 billion. Called Stargate, the aim would be to have it operational by 2008 to drive OpenAI’s next generation of artificial intelligence. According to reports, the Stargate complex would span hundreds of U.S. acres and use up to 5 gigawatts of power, which is massive (the equivalent of a substantial metropolitan power grid). In light of those power needs, a nuclear power source is said to be under consideration. The project is not yet green-lit, and no U.S. location has been selected. Continue reading Microsoft, OpenAI Considering a Supercomputer Data Center

Databricks DBRX Model Offers High Performance at Low Cost

Databricks, a San Francisco-based company focused on cloud data and artificial intelligence, has released a generative AI model called DBRX that it says sets new standards for performance and efficiency in the open source category. The mixture-of-experts (MoE) architecture contains 132 billion parameters and was pre-trained on 12T tokens of text and code data. Databricks says it provides the open community and enterprises who want to build their own LLMs with capabilities previously limited to closed model APIs. Compared to other open models, Databricks claims it outperforms alternatives including Llama 2-70B and Mixtral on certain benchmarks. Continue reading Databricks DBRX Model Offers High Performance at Low Cost

Deepgram’s Speech Portfolio Now Includes Human-Like Aura

Deepgram’s new Aura software turns text into generative audio with a “human-like voice.” The 9-year-old voice recognition company has raised nearly $86 million to date on the strength of its Voice AI platform. Aura is an extremely low-latency text-to-speech voice AI that can be used for voice AI agents, the company says. Paired with Deepgram’s Nova-2 speech-to-text API, developers can use it to “easily (and quickly) exchange real-time information between humans and LLMs to build responsive, high-throughput AI agents and conversational AI applications,” according to Deepgram. Continue reading Deepgram’s Speech Portfolio Now Includes Human-Like Aura

GTC: Nvidia Unveils Blackwell GPU for Trillion-Parameter LLMs

Nvidia unveiled what it is calling the world’s most powerful AI processing system, the Blackwell GPU, purpose built to power real-time generative AI on trillion-parameter large language models at what the company says will be up to 25x less cost and energy consumption than its predecessors. Blackwell’s capabilities will usher in what the company promises will be a new era in generative AI computing. News from Nvidia’s GTC 2024 developer conference included the NIM software platform, purpose built to streamline the setup of custom and pre-trained AI models in a production environment, and the DGX SuperPOD server, powered by Blackwell. Continue reading GTC: Nvidia Unveils Blackwell GPU for Trillion-Parameter LLMs

Apple Unveils Progress in Multimodal Large Language Models

Apple researchers have gone public with new multimodal methods for training large language models using both text and images. The results are said to enable AI systems that are more powerful and flexible, which could have significant ramifications for future Apple products. These new models, which Apple calls MM1, support up to 30 billion parameters. The researchers identify multimodal large language models (MLLMs) as “the next frontier in foundation models,” which exceed the performance of LLMs and “excel at tasks like image captioning, visual question answering and natural language inference.” Continue reading Apple Unveils Progress in Multimodal Large Language Models

Grok-1 Architecture Open-Sourced for General Release by xAI

Elon Musk’s xAI has released its Grok chatbot and open-sourced part of the underlying Grok-1 model architecture for any developer or entrepreneur to use for purposes including commercial applications. Musk unveiled Grok in November and announced that it would be publicly released this month. The chatbot itself is available to X social premium members, who can ask the cheeky AI questions and get answers with a snarky attitude inspired by “The Hitchhiker’s Guide to the Galaxy” sci-fi novel. The training for Grok’s foundation LLM is said to include X social posts. Continue reading Grok-1 Architecture Open-Sourced for General Release by xAI

Meta Building Giant AI Model to Power Entire Video Ecosystem

Facebook chief Tom Alison says parent company Meta Platforms is building a giant AI model that will eventually “power our entire video ecosystem.” Speaking at the Morgan Stanley Technology, Media & Telecom Conference this week, Alison said the model will drive the company’s video recommendation engine across all platforms that host long-form video as well as the short-form Reels, which are limited to 90 seconds. Alison said the company began experimenting with the new, super-sized AI model last year and found that it helped improve Facebook’s Reels watch time by anywhere from 8-10 percent. Continue reading Meta Building Giant AI Model to Power Entire Video Ecosystem

Anthropic’s Claude 3 AI Is Said to Have ‘Near-Human’ Abilities

Anthropic has released Claude 3, claiming new industry benchmarks that see the family of three new large language models approaching “near-human” cognitive capability in some instances. Accessible via Anthropic’s website, the three new models — Claude 3 Haiku, Claude 3 Sonnet and Claude 3 Opus — represent successively increased complexity and parameter count. Sonnet is powering the current Claude.ai chatbot and is free, for now, requiring only an email sign-in. Opus comes with the the $20 monthly subscription for Claude Pro. Both are generally available from the Anthropic website and via API in 159 countries, with Haiku coming soon. Continue reading Anthropic’s Claude 3 AI Is Said to Have ‘Near-Human’ Abilities

France’s Mistral AI Makes Its Global Debut on Microsoft Azure

Paris-based startup Mistral AI has made an immediate splash in the world of artificial intelligence, securing partnerships with IBM, Microsoft and others nine months after its launch. The company is offering natural language processing models, including its flagship Mistral Large, which becomes only the second LLM (after OpenAI) to land a commercial berth on Microsoft’s Azure cloud, where Meta Platforms’ Llama 2 is available in preview. Boasting “top-tier reasoning capacities” and sophisticated conversational capabilities, Mistral Large specializes in “reasoning, analysis and generation (RAG), is multilingual and supports up to 32,000 tokens.” Continue reading France’s Mistral AI Makes Its Global Debut on Microsoft Azure

MWC: Qualcomm Unveils AI Hub and Promotes 5G, 6G Tech

Qualcomm raised the curtain on a variety of artificial intelligence, 5G, and Wi-Fi technologies at Mobile World Congress Barcelona, which runs through Thursday. The San Diego-based chip designer unveiled an AI Hub it says will help developers create voice-, text- and image-based applications using pre-optimized AI models. Qualcomm’s flagship AI chips — the mobile Snapdragon 8 Gen 3 processor and the PC-centric Snapdragon X Elite — were announced last year. With the first splash of products now heading to market the company is promising to push the boundaries of 5G and 6G. Continue reading MWC: Qualcomm Unveils AI Hub and Promotes 5G, 6G Tech

Reddit Announces IPO on Heels of Expanded Deal with Google

Community message board and social news aggregator Reddit, founded in 2005, has filed to go public on the New York Stock Exchange in an IPO observers say may be complete in a matter of weeks. It is the first social media company to go public in many years, with Snap Inc.’s 2017 offering cited as the most recent stock market splash. Reddit’s bankers are reportedly seeking a $5 billion valuation, about half the $10 billion it was valued at for a 2021 private funding round. Reddit filed with the SEC the same day it announced an “expanded partnership” with Google to use Vertex AI. Continue reading Reddit Announces IPO on Heels of Expanded Deal with Google