Applied Research
At CHAI, all of our growth is driven by better AI.

Optimize
CHAI's research focus is on applying proven techniques such as RLHF, SFT, Prompt Engineering, Rejection Sampling, and LLM routing. We iterate across various algorithms such as DPO, LoRA, training reward models, and embeddings to build recommender systems. Only by making the AI better do we see an increase in metrics such as monetization and engagement.
Scale
It is well-known that scaling up LLMs improves their performance.
There are many dimensions to this scale. For example: parameters,
dataset size, inference compute, context length, and the number of
LLMs served. This creates an engineering challenge.
Firstly, scaling up tends to increase costs; this drives us to
experiment with techniques such as quantization, custom CUDA
kernels, Flash Attention and KV-Caching. Secondly, at a certain
scale, out-of-the-box solutions tend to breakdown. This drives us
to build our own custom implementations such as our own
self-managed Kubernetes cluster or inference engines.

Talks and articles
Blog
Feb 04, 2025
Talk
Jan 28, 2025
Talk
Nov 20, 2024
Blog
Apr 01, 2024
Blog
Feb 13, 2024
Talk
Jan 09, 2024
Paper
Jan 04, 2024
Blog
Nov 08, 2023
talk
Oct 04, 2023
talk
Oct 04, 2023
Paper
Mar 30, 2023