ByteDance Scales Offline Inference with Multi-Modal LLMs to 200 TB Data
2023-09-04
ThirdAI Uses Ray for Parallel Training of Billion-Parameter NN on Commodity CPUs
2023-08-30
Llama 2 is about as factually accurate as GPT-4 for summaries and is 30X cheaper
2023-08-29
Fine Tuning is for form not facts
2023-08-27
Llama 2 is about as factually accurate as GPT-4 for summaries and is 30X cheaper
2023-08-23
ByteDance Scales Offline Inference with Multi-Modal LLMs to 200 TB Data
2023-08-15
Continuous batching to increase LLM inference throughput and reduce p50 latency
2023-08-15
Design considerations for RAG application in production mode
2023-08-14
Numbers every LLM Developer should know
2023-08-12
Fine-Tuning Llama-2: A Comprehensive Case Study for Tailoring Custom Models
2023-08-11