一区二区日本_久久久久久久国产精品_无码国模国产在线观看_久久99深爱久久99精品_亚洲一区二区三区四区五区午夜_日本在线观看一区二区

Strategic Round II
Complete

READ MORE
Chai AI logo

CHAI: Chat + AI

Quant traders building AI Platform
Palo Alto, CA

[ Daily Active Users Growth ]

Incentives & Scale

RESEARCH

All platforms work best with the right incentives. At CHAI, we've tried paying developers, but the biggest motivators remain high-quality feedback, recognition, and the satisfaction of building a popular LLM. Our scale enables the critical mass of feedback and models needed to create strong feedback loops.

Graph showing Chai DAU growth
OCT 2022
APR 2023
OCT 2023
APR 2024
OCT 2024
APR 2025
NOV 2022

CHAI Launches on App Store

We were the first to launch a consumer AI platform, allowing users to create their own ChatAIs—ahead of Character AI and ChatGPT.

FEB 2023

Deploys First In-House 6B LLM

Open-sourced LLMs no longer satisfied our users' requirements, as the LLMs needed to be adapted for social and engagement purposes. We saw a +10% engagement boost from our own in-house model.

MAR 2023

Deploys Best-of-4 Reward Model

We continued to iterate on RLHF (Reinforcement Learning with Human Feedback), training a reward model directly on user signals. This led to a huge boost in our day 30 user retention.

APR 2023

Larger Model Upgrade - 13B Architecture

We found that a bigger model leads to better depth, therefore better retention. We re-trained our LLM from scratch and saw another +10% engagement boost.

MAY 2023

PPO Model Deployed

Using Proximal Policy Optimization, a reinforcement learning technique, we optimized our base foundation model to decrease the probability a chat session ends.

JUNE 2023

Deploys Reward Model XL

Continued to scale up our reward model. Trained with 100 million signals to decrease user retry rate and increase chat session length.

OCT 2023

Efficient Inference & Custom GPU Orchestration

Off-the-shelf load balancing and vLLM were no longer sufficient to support our user base at 500K DAU scale. We implemented custom CUDA kernels together with our own GPU orchestration system.

NOV 2023

Increased GPU Reservation

We hit a scaling issue due to high demand from our users. We reserved an additional 1,000 A100 GPUs from our provider to scale reliably.

NOV 2023

Deployed Model Blending

CHAI invented model blending—ensembling different LLMs trained on different targets at the conversation level. This outperformed GPT-3 in user retention.

DEC 2023

BO8 Reward Model Deployed

With increased cluster capacity, we implemented Best-of-8 rejection sampling, utilizing our upgraded reward model to its full extent.

MAR 2024

DPO Model Deployed

Utilizing Direct Preference Optimization with user preference datasets, we boosted engagement by 20%. The performance stacked well with our existing reward model.

AUG 2024

Upgraded All Existing Blends to DPO

Building on the success of DPO, we iterated on optimization targets and data selection, and successfully deployed DPO across all production blends.

SEP 2024

13B Reward Model Deployed

With increased GPU capacity due to cluster upgrades, we were able to serve larger reward models for all users.

OCT 2024

10x 24B Models Deployed

We upgraded our existing production blend to 24B models. With blending enabled, we saw a surge in daily active users and day 30 retention.

JAN 2025

Model Mesh Orchestrator Deployed

To support over 1M Daily Active Users, Model Mesh—an in-house cluster orchestration platform—was deployed to handle multi-cluster, multi-GPU-type serving of hundreds of LLMs in production.

MAR 2025

GRPO Deployed

GRPO (Group Relative Policy Optimization) is an upgrade from Direct Preference Optimization, resulting in a +15% engagement improvement.

[ Product ]

Building Platform for Social AI

We believe in platforms. There is huge demand for AI that is not only factually correct but also entertaining and social.

Gradient background Gradient background Gradient background
IOS ANDROID
[ GPU Cluster ]

1.4 EXAFLOPS GPU CLUSTER
FOR AI INFERENCE

CLUSTER

At CHAI, we serve hundreds of in-house trained LLMs across several GPU chip types from both AMD and Nvidia. While open-source solutions such as vLLM work well for simple workloads, we've found that we can further improve upon vLLM by almost an order of magnitude through several optimizations, such as custom kernels and compute-efficient attention approximations.

NUMBER OF GPUS
5000 GPUs
NUMBER OF TOKENS SERVED
1.2T Tokens / Day
NUMBER OF UNIQUE LLMS SERVED
51K LLMs
CLUSTER COMPUTE PERFORMANCE
>1.4 Exaflops
NVIDIA A100
NVIDIA A100
NVIDIA L40S
NVIDIA L40S
AMD Mi325x
AMD Mi325x
AMD Mi300x
AMD Mi300x

Current openings

JOBS

Who we are

NEWS
主站蜘蛛池模板: 亚洲精品中文字幕乱码三区91 | 少妇福利视频 | 日韩欧美一区二区三区四区 | 久久夜色精品国产欧美乱极品 | 日韩性生活视频 | 欧美二区视频 | 日韩精品一区二区视频 | 日皮视频免费看 | 韩日中文字幕 | 国产一区二区三区视频 | 久久h| 日韩av免费在线 | 国产日韩欧美一区 | 国产精品久久久久久妇女6080 | av观看免费 | 久久综合久久鬼 | 第一福利视频导航 | 蜜桃视频一区 | 在线观看免费黄色 | 欧美视频在线观看一区 | 日日干夜夜草 | 国产三级在线 | 久色精品 | 日韩一二三 | 中文在线免费观看 | 日韩视频网 | 日韩一区二区三区四区 | 国产日韩综合 | 国产日韩欧美一区二区 | 久久在线免费视频 | 91国内精品| 中文字幕一区在线 | 在线色| 国产免费一级 | 黄色网在线 | 免费看黄色网址 | 国产成人精品一区二区三区在线观看 | 国产成人精品免费视频 | 欧美 日韩 国产 成人 在线 | 一级黄视频| 色在线视频 |