Large language model
Researchers from Huawei Noah’s Ark Lab and Peking University, in collaboration with Huawei Consumer Business Group, have developed PanGu-π Pro, a groundbreaking tiny language model for mobile devices. The model achieves high performance through strategic optimization, compression of the tokenizer, and architectural adjustments, setting new benchmarks for compact language models. This innovation opens new avenues…
Hydragen is a transformative solution in optimizing large language models (LLMs). Developed by research teams from Stanford University, the University of Oxford, and the University of Waterloo, Hydragen’s innovative attention decomposition method significantly enhances computational efficiency for shared-prefix scenarios, showcasing up to a 32x improvement in LLM throughput and adaptable application to various settings. For…
OpenAI’s innovative text-to-video model, Sora, is transforming digital content creation. It offers unparalleled capabilities to generate, extend, and animate high-quality videos with remarkable detail. By leveraging spacetime patches and recaptioning techniques, Sora demonstrates diverse applications, showcasing potential for AGI and simulating real-world dynamics. Despite limitations, Sora represents a significant leap forward in AI-driven video generation.
AI development is evolving from static, task-centric models to dynamic, adaptable agent-based systems suitable for various applications. Recent research proposes the Interactive Agent Foundation Model, a multi-modal system with unified pre-training to process text, visual data, and actions. It demonstrates promising efficacy across diverse domains, showing potential for generalist agents in AI advancement.
The Nomic AI’s nomicembed-text-v1 model revolutionizes long-context text embeddings, boasting a sequence length of 8192, surpassing predecessors in performance evaluations. Open-source with an Apache-2 license, it emphasizes transparency and accessibility, setting new AI community standards. Its development process prioritizes auditability and potential replication, heralding a future of profound understanding in human discourse.
Researchers from Fudan University, Ohio State University, and Pennsylvania State University, Meta AI, have developed TravelPlanner, an AI benchmark to evaluate agents’ planning skills in realistic scenarios. It challenges AI agents to plan multi-day travel itineraries, highlighting limitations in current AI models. TravelPlanner aims to advance AI planning capabilities and bridge the gap between theoretical…
MeetKai, an influential player in conversational AI, introduced Functionary, an open-source language model for function calling. In contrast to larger models like GPT-4, Functionary offers faster, more cost-effective inference with high accuracy. It seamlessly integrates with OpenAI’s platform and aligns with MeetKai’s vision for the metaverse, inviting developers to shape the future of applied generative…
LMMs have widely expanded using CLIP for vision encoding and LLMs for multi-modality reasoning. Scaling up CLIP is crucial, leading to the EVA-CLIP-18B model with 18B parameters. It achieves remarkable zero-shot top-1 accuracy on 27 benchmarks and demonstrates effectiveness in various image tasks, underlining progress in open-source AI models. [50 words]
Graph Neural Networks (GNNs) leverage graph structures to perform inference on complex data, addressing the limitations of traditional ML algorithms. Google’s TensorFlow GNN 1.0 (TF-GNN) library integrates with TensorFlow, enabling scalable training of GNNs on heterogeneous graphs. It supports supervised and unsupervised training, subgraph sampling, and flexible model building for diverse tasks.
Vision Language Models (VLMs) leverage Large Language Models’ strength to comprehend visual data, demonstrating capability in visual question answering and optical character recognition. A study by Tsinghua University and Zhipu AI introduces Chain of Manipulations (CoM) to enable VLMs for visual reasoning, leading to competitive performance on various benchmarks and highlighting potential for accelerated VLM…
DeepSeekMath, developed by DeepSeek-AI, Tsinghua University, and Peking University, revolutionizes mathematical reasoning using large language models. With a dataset of over 120 billion tokens of math-related content and innovative training using Group Relative Policy Optimization, it achieves a top-1 accuracy of 51.7% on the MATH benchmark, setting a new standard for AI-driven mathematics.
State-space models (SSMs) are being explored as an alternative to Transformer networks in AI research. SSMs aim to address computational inefficiencies in Transformer networks and have led to the proposal of MambaFormer, a hybrid model combining SSMs and Transformer attention blocks. MambaFormer demonstrates superior in-context learning capabilities, offering new potential for AI advancement.
Large Language Models, like GPT-3, have revolutionized Natural Language Processing by scaling to billions of parameters and incorporating extensive datasets. Researchers have also introduced Speech Language Models directly trained on speech, leading to the development of SPIRIT-LM. This multimodal language model seamlessly integrates text and speech, demonstrating potential impacts on various applications.
The introduction of Large Language Models in Artificial Intelligence, propelled by the transformer architecture, has greatly enhanced machines’ ability to comprehend and solve problems akin to human cognition. USC and Google’s researchers have introduced SELF-DISCOVER, improving these models’ reasoning capabilities significantly, bridging the gap between Artificial Intelligence and human cognitive processes.
OpenMoE revolutionizes Natural Language Processing (NLP) with its Mixture-of-Experts approach, scaling model parameters efficiently for enhanced task performance. OpenMoE’s comprehensive suite of decoder-only LLMs, meticulously trained on extensive datasets, showcases commendable cost-effectiveness and competitive performance. Moreover, the project’s open-source ethos democratizes NLP research, establishing a new standard for future LLM development.
Researchers have developed a regression-based deep-learning method, CAMIL, to predict continuous biomarkers from pathology slides, surpassing classification-based methods. The approach significantly improves prediction accuracy and aligns better with clinically relevant regions, particularly in predicting HRD status. This advancement demonstrates the potential of regression models in enhancing prognostic capabilities in digital pathology. Further research is recommended…
This text discusses the problematic behaviors exhibited by language models (LMs) and proposes strategies to enhance their robustness. It emphasizes automated adversarial testing techniques to identify vulnerabilities and elicit undesirable behaviors. Researchers at Eleuther AI focus on identifying well-formed language prompts to elicit arbitrary behaviors while maintaining naturalness. They introduce reverse language modeling to optimize…
Artificial Intelligence (AI) has seen significant advancements in the past decade, with generative AI posing security and privacy threats due to its ability to create realistic content. Meta’s AudioSeal is a novel audio watermarking technique designed to detect and localize AI-generated speech, outperforming previous methods in speed and accuracy. [49 words]
The study introduces LEAP, a method that incorporates mistakes into AI learning. It improves model reasoning abilities and performance across tasks like question answering and mathematical problem-solving. This approach is significant for its potential to make AI models more adaptable and intelligent, akin to human learning processes. LEAP marks a significant step towards more intelligent…
Large-scale training of generative models on video and image data is explored, utilizing text-conditional diffusion models. A transformer architecture operates on video and image latent codes to enable generation of high-fidelity video. Sora, the largest model, can generate a minute of video. Scaling video generation models shows promise for building general purpose simulators of the…