Llama 3.1 Megathread
Llama 3.1 Megathread
Llama 3.1 Megathread
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Develop Alongside Local LLMs w/ Open Interpreter
What open-source LLMs are you using in 2024?
FOSAI 2024
Blaed's Hiatus (Part I)
What kind of content do you want to see more of?
Llama 2 / WizardLM Megathread
Sharing brev.dev - A new platform for fine-tuning models on cloud GPUs
Create a Large Language Model from Scratch with Python – Tutorial
HyperTech News Report #0003 - Expanding Horizons
HyperTech News Report #0003 - Expanding Horizons
HyperTech News Report #0003 - Expanding Horizons
Anyone else working with retrieval augmented generation? (RAG)
Mistral 7B Megathread
AutoGen - Enabling Next Generation LLM Applications
What do you think are some of the most interesting use cases for AGI?
Why do you like LLMs?
What sort of tokens per second are you seeing with your hardware? Mind sharing some notes on what you're running there? Super curious!