ruliad

Login

27 . 11 . 2024

Introducing DeepThought-8B: A small, capable reasoning model

Today we're releasing DeepThought-8B, a small, capable AI reasoning model built on LLaMA-3.1 8B. This release represents our first step toward making AI reasoning more transparent and controllable, while demonstrating that smaller, more efficient models can achieve sophisticated reasoning capabilities that rival models of much larger scales. DeepThought-8B unlocks test-time compute scaling during inference for all- taking as many reasoning steps as needed to solve complex problems.

We're excited to make DeepThought-8B available through our chat application today, with powerful features that allow you to modulate the way the model reasons. In the coming weeks, we will be opening up our developer API (currently in closed BETA) and regular updates of the open-source model weights.


What makes DeepThought different?

DeepThought-8B approaches tasks differently from traditional language models. Given a problem, it breaks down its thinking into clear steps until it reaches a conclusion. Here's what this looks like in practice:

{
  "step": 1,
  "type": "problem_understanding",
  "thought": "The user is asking how many Rs there are in the word 'strawberry'"
}


Each step is documented, making it easier to understand how the model arrives at its answers.


Key Features

  1. Transparent Reasoning: The model shows its work, step by step.

  2. Programmable Approach: You can guide how the model reasons without retraining it through our API.

  3. Test-time Computer: The model can take as many steps as needed to solve a problem.

  4. Small but Mighty: At 8B parameters, DeepThought runs on consumer GPUs with 16GB+ VRAM, making sophisticated AI reasoning accessible without requiring enterprise-grade hardware.

  5. Structured Output: Consistent JSON-formatted reasoning chains for easy integration.



Performance

While we're excited about DeepThought-8B's capabilities, we believe in transparency and community-driven evaluation. Our internal testing shows promising results across reasoning, math and coding benchmarks, but we encourage you to test these capabilities yourself.



Some early findings

  • Strong performance in step-by-step problem-solving.

  • Competitive results on coding and mathematical tasks

  • Reliable instruction following with transparent reasoning chains

  • Performance scales with test-time compute, allowing deeper reasoning on complex tasks.
  • However, like all models, DeepThought-8B has its limitations. We're actively working on:

  • Improving mathematical reasoning for complex problems.

  • Enhancing long-context processing.

  • Increasing competence in edge cases


  • Rather than hyping benchmark scores that might not reflect real-world usage, we invite you to:

    1. Test the model in your specific use cases.

    2. Share your findings with our community.

    3. Help us identify areas for improvement.


    You can report your findings and discuss the model's performance by tagging or DMing @ruliad_ai on X. You can also send us an email at team@ruliad.co.

    Try DeepThought now by going tochat.ruliad.co.

    Download DeepThought-8B-Llama-v0.01-alpha now by going toour huggingface repo.

    What's next? This is just the beginning. We'll be regularly updating the model based on your feedback and our ongoing research.

    ruliad

    Self-Sovereign Intelligence for all

    © 2024 Ruliad Technologies. All Rights Reserved.