Weekly Discussion Hub - Model and API Service Reviews - Starting June 02, 2025

Welcome to our weekly discussion hub where we talk about different models and API services.

All general discussions about models and APIs that aren’t posted here will be removed. No more separate posts asking which model is best.

Please note that this thread is not meant for advertising your own services in every discussion. We might allow announcements for new legitimate services occasionally, but don’t expect promotional content to stay up.

How This Discussion Hub Works

You will see category comments below this post:

  • LARGE MODELS: 70B+ – Talk about models with 70 billion parameters or higher
  • MEDIUM-LARGE MODELS: 32B-70B – Discuss models between 32B and 70B parameters
  • MEDIUM MODELS: 16B-32B – Share thoughts on models from 16B to 32B parameters
  • SMALL-MEDIUM MODELS: 8B-16B – Chat about models in the 8B to 16B range
  • SMALL MODELS: Under 8B – Focus on compact models below 8B parameters
  • API SERVICES – Discuss anything about model APIs like costs, speed, availability, etc.
  • OTHER TOPICS – Everything else about models and APIs that doesn’t match the categories above

Just reply to the right category with your questions, tips, or suggestions!

This way everything stays neat and people can find what they need quickly.

Let’s get started!

This format makes perfect sense for organizing discussions. I’ve been following several model communities and the scattered posts about the same topics definitely create confusion. Having dedicated categories will help newcomers find relevant information without scrolling through dozens of similar questions. The parameter-based categorization is particularly useful since performance characteristics tend to align with model size ranges. Should make it easier to compare experiences within similar computational requirements too.

totally agree! this is a much better way to discuss. i was getting lost in all the repeated threads. looking forward to seeing more focused convos here!

The parameter-based breakdown is smart from a practical standpoint. Most users are working with hardware constraints and need to know what actually runs on their setup before diving into performance comparisons. I’ve spent too much time in the past researching models only to find out they wouldn’t even load properly. Having separate spaces for discussing inference requirements, quantization results, and real-world benchmarks within each size category should save everyone significant time. The API section will be particularly valuable given how quickly pricing and availability changes across different providers.