-
Amazon Researchers Present a Deep Learning Compiler for Training Consisting of Three Main Features- a Syncfree Optimizer, Compiler Caching, and Multi-Threaded Execution
A team of researchers has developed a deep learning compiler for neural network training. The compiler includes a sync-free optimizer, compiler caching, and multi-threaded execution, resulting in significant speedups and resource efficiency compared to traditional approaches. The compiler improves training procedures for real-world applications and has the potential to optimize neural network models across various…
-
Optimize pet profiles for Purina’s Petfinder application using Amazon Rekognition Custom Labels and AWS Step Functions
Purina US, a subsidiary of Nestle, used artificial intelligence (AI) and machine learning (ML) to automate animal breed detection on the Petfinder platform. By leveraging Amazon Rekognition Custom Labels, AWS Step Functions, and other AWS services, Purina created an ML model that detects the pet breed from uploaded images and auto-populates pet attributes. This solution…
-
This AI Research Introduces Flash-Decoding: A New Artificial Intelligence Approach Based on FlashAttention to Make Long-Context LLM Inference Up to 8x Faster
Flash-Decoding is a groundbreaking technique that improves the efficiency of large language models during the decoding process. It addresses the challenges associated with attention operation, making the models up to 8 times faster. By optimizing GPU utilization, Flash-Decoding reduces operational costs and promotes greater accessibility of these models in various applications. This innovation is a…
-
21-Year-Old Student Deciphered of Ancient Herculaneum Scrolls Using AI
21-year-old Luke Farritor, a computer science student at the University of Nebraska-Lincoln, has made a groundbreaking discovery by using a machine-learning algorithm to read the first-ever text from a burnt scroll found in the ancient city of Herculaneum. His breakthrough could lead to the deciphering of numerous currently unreadable ancient texts. Farritor won $40,000 in…
-
China has a new plan for judging the safety of generative AI—and it’s packed with details
China’s National Information Security Standardization Technical Committee has released a draft document outlining rules for determining problematic generative AI models. The document provides criteria for banning data sources, demands diversification of training materials, and sets requirements for hiring moderators. It also outlines what constitutes prohibited content and addresses the need for more subtle censorship. While…
-
This AI Research Presents RoboHive: A Comprehensive Software Platform and Ecosystem for Research in the Field of Robot Learning and Embodied Artificial Intelligence
Researchers have developed RoboHive, a platform for robot learning, to address the challenges in this field. RoboHive serves as a benchmarking and research tool, offering various learning paradigms and hardware integration. Its key features include a wide range of contexts, teleoperation support, visual diversity, clear metrics, and baseline results. The goal is to bridge the…
-
Nvidia and Foxconn to build ‘AI factory’ to make EVs
Nvidia and Foxconn are joining forces to build “AI factories” that will accelerate the production of autonomous electric vehicles (EVs). Foxconn, known for manufacturing Apple’s iPhone, aims to capture 5% of the EV manufacturing market by 2025. The factories will incorporate cutting-edge manufacturing and AI systems to develop and improve EVs. Nvidia’s technologies, including Drive…
-
Microsoft Azure AI Introduces Idea2Img: A Self-Refinancing Multimodal AI Framework For The Development And Design Of Images Automatically
Microsoft Azure AI has developed Idea2Img, a self-refinancing multimodal framework for automated image design and generation. Idea2Img utilizes a large language model (GPT-4V) and a text-to-image model to iterate and refine image creation based on user input. The framework demonstrates improved semantic and visual quality in image generation, outperforming other models in user preference studies.
-
Run Zephyr 7B with an API
Zephyr 7B alpha outperforms Llama 2 70B Chat on MT Bench. Simple code lines teach you how to run it efficiently.
-
Researchers from NVIDIA Introduce Retro 48B: The Largest LLM Pretrained with Retrieval before Instruction Tuning
Researchers from Nvidia and the University of Illinois at Urbana-Champaign have developed Retro 48B, a larger language model that improves on previous retrieval-augmented models. By pre-training with retrieval on a vast corpus, Retro 48B enhances task performance in question answering. The study demonstrates the potential of larger retrieval-augmented models in natural language understanding.