-
DAI#24 – Brain chips, clones, and Swifties fight back
This week’s AI news features the following highlights: 1. Taylor Swift’s battle against explicit AI deep fake images and the concerning ease of generating such content using AI. 2. The rise of political deep fakes showcasing AI’s capabilities in replicating voices with convincing realism and the challenges of detecting these fakes. 3. OpenAI’s evolving transparency…
-
Meet CMMMU: A New Chinese Massive Multi-Discipline Multimodal Understanding Benchmark Designed to Evaluate Large Multimodal Models LMMs
The CMMMU benchmark has been introduced to bridge the gap between powerful Large Multimodal Models (LMMs) and expert-level artificial intelligence in tasks involving complex perception and reasoning with domain-specific knowledge. It comprises 12,000 Chinese multimodal questions across six core disciplines and employs a rigorous data collection and quality control process. The benchmark evaluates LMMs, presents…
-
DeepSeek-AI Introduce the DeepSeek-Coder Series: A Range of Open-Source Code Models from 1.3B to 33B and Trained from Scratch on 2T Tokens
The integration of large language models (LLMs) in software development has revolutionized code intelligence, automating aspects of programming and increasing productivity. Disparities between open-source and closed-source models have hindered accessibility and democratization of advanced coding tools. DeepSeek-AI and Peking University’s DeepSeek-Coder series addresses this gap, enhancing open-source models’ functionality and performance, marking a significant advancement…
-
This AI Paper from China Introduces ‘AGENTBOARD’: An Open-Source Evaluation Framework Tailored to Analytical Evaluation of Multi-Turn LLM Agents
AgentBoard, developed by researchers from multiple Chinese universities, presents a benchmark framework and toolkit for evaluating LLM agents. It addresses challenges in assessing multi-round interactions and diverse scenarios in agent tasks. With a fine-grained progress rate metric and interactive visualization, it illuminates the capabilities and limitations of LLM agents across varied environments.
-
Researchers from the Chinese University of Hong Kong and Tencent AI Lab Propose a Multimodal Pathway to Improve Transformers with Irrelevant Data from Other Modalities
The researchers from The Chinese University of Hong Kong and Tencent AI Lab introduce the Multimodal Pathway Transformer (M2PT) to enhance transformer performance by incorporating irrelevant data from other modalities, resulting in substantial performance improvements across various recognition tasks. The approach involves Cross-Modal Re-parameterization and demonstrates tangible implementation of auxiliary weights without incurring inference costs.
-
Data center energy demands are outstripping what the grid can provide
The demand for AI is challenging environmental sustainability, as it significantly increases electricity consumption. Data centers, particularly those supporting generative AI, strain global energy infrastructure. The rising electricity demands from AI and data centers are creating environmental and grid stability concerns, urging the need for more sustainable practices within the AI industry and alternative energy…
-
Meet BiTA: An Innovative AI Method Expediting LLMs via Streamlined Semi-Autoregressive Generation and Draft Verification
Recent advancements in large language models (LLMs) like Chat-GPT and LLaMA-2 have led to an exponential increase in parameters, posing challenges in inference delay. To address this, Intellifusion Inc. and Harbin Institute of Technology propose Bi-directional Tuning for lossless Acceleration (BiTA) to expedite LLMs, achieving significant speedups without compromising output quality. (50 words)
-
A New Research Study from the University of Surrey Shows Artificial Intelligence Could Help Power Plants Capture Carbon Ising 36% Less Energy from the Grid
Researchers from the University of Surrey have used AI to improve carbon capture technology. By employing AI algorithms, they achieved a 16.7% increase in CO2 capture and reduced energy usage by 36.3%. The system employed packed bubble column reactor and machine learning techniques to optimize performance. This study demonstrates the potential of AI in creating…
-
UC Berkeley and UCSF Researchers Propose Cross-Attention Masked Autoencoders (CrossMAE): A Leap in Efficient Visual Data Processing
Researchers from UC Berkeley and UCSF have introduced Cross-Attention Masked Autoencoders (CrossMAE) in computer vision, aiming to enhance processing efficiency for visual data. By leveraging cross-attention exclusively for decoding masked patches, CrossMAE simplifies and expedites the decoding process, achieving substantial computational reduction while maintaining quality and performance in complex tasks. This research presents a groundbreaking…
-
OpenAI says GPT-4 could help you make a bioweapon, maybe
RAND and OpenAI issued conflicting reports on the possibility of using AI for bioweapon development. OpenAI’s study, involving biology experts and internet access, found that access to a research version of GPT-4 may enhance the ability to access biological threat information but emphasized that information access alone is insufficient for bioweapon creation. The study concluded…