OpenAI

1602 Articles
Layer Parallelism: Enhancing LLM Inference Efficiency Through Parallel Execution of Transformer Layers
OpenAI

Layer Parallelism: Enhancing LLM Inference Efficiency Through Parallel Execution of Transformer Layers

LLMs have demonstrated exceptional capabilities, but their substantial computational demands pose significant challenges for large-scale deployment. While previous studies indicate that intermediate layers...

Open O1: Revolutionizing Open-Source AI with Cutting-Edge Reasoning and Performance
OpenAI

Open O1: Revolutionizing Open-Source AI with Cutting-Edge Reasoning and Performance

The Open O1 project is a groundbreaking initiative aimed at matching the powerful capabilities of proprietary models, particularly OpenAI’s O1, through an open-source...

Step by Step Guide on How to Build an AI News Summarizer Using Streamlit, Groq and Tavily
OpenAI

Step by Step Guide on How to Build an AI News Summarizer Using Streamlit, Groq and Tavily

Introduction In this tutorial, we will build an advanced AI-powered news agent that can search the web for the latest news on a...

Google DeepMind Research Introduces WebLI-100B: Scaling Vision-Language Pretraining to 100 Billion Examples for Cultural Diversity and Multilingualit
OpenAI

Google DeepMind Research Introduces WebLI-100B: Scaling Vision-Language Pretraining to 100 Billion Examples for Cultural Diversity and Multilingualit

Machines learn to connect images and text by training on large datasets, where more data helps models recognize patterns and improve accuracy. Vision-language...

Can Users Fix AI Bias? Exploring User-Driven Value Alignment in AI Companions
OpenAI

Can Users Fix AI Bias? Exploring User-Driven Value Alignment in AI Companions

Large language model (LLM)–based AI companions have evolved from simple chatbots into entities that users perceive as friends, partners, or even family members....

Can 1B LLM Surpass 405B LLM? Optimizing Computation for Small LLMs to Outperform Larger Models
OpenAI

Can 1B LLM Surpass 405B LLM? Optimizing Computation for Small LLMs to Outperform Larger Models

Test-Time Scaling (TTS) is a crucial technique for enhancing the performance of LLMs by leveraging additional computational resources during inference. Despite its potential,...

Meta AI Introduces CoCoMix: A Pretraining Framework Integrating Token Prediction with Continuous Concepts
OpenAI

Meta AI Introduces CoCoMix: A Pretraining Framework Integrating Token Prediction with Continuous Concepts

The dominant approach to pretraining large language models (LLMs) relies on next-token prediction, which has proven effective in capturing linguistic patterns. However, this...

Anthropic AI Launches the Anthropic Economic Index: A Data-Driven Look at AI’s Economic Role
OpenAI

Anthropic AI Launches the Anthropic Economic Index: A Data-Driven Look at AI’s Economic Role

Artificial Intelligence is increasingly integrated into various sectors, yet there is limited empirical evidence on its real-world application across industries. Traditional research methods—such...