IBM joins the party of reasoning models with Granite 3.2 | Large language model certification | Top llm models huggingface | Popular language models | Turtles AI

IBM joins the party of reasoning models with Granite 3.2
Preview reveals advanced techniques to improve accuracy and flexibility in AI responses, balancing efficiency and safety
Isabella V9 February 2025

 

IBM previews new reasoning capabilities in the Granite family of language models, improving the quality of responses through advanced processing at the inference stage. This approach allows for improved response accuracy without compromising efficiency and security.

Key Points:

  • New Reasoning Capabilities in Granite Models: IBM develops advanced reasoning techniques without sacrificing efficiency and security.
  • Reinforcement learning approach: The Granite-3.1-8b-instruct model autonomously improves its reasoning capabilities.
  • Flexible reasoning activation: Developers can modulate the level of processing to balance the quality and speed of responses.
  • Comparison with other advanced models: Granite offers a better balance of accuracy, versatility, and security than competing models.

IBM is redefining the concept of reasoning in AI models with a new preview of Granite’s advanced capabilities. The goal is to provide greater processing power during inference, significantly improving the quality of answers through logically splitting and deep analysis of the problems addressed. This advancement builds on the concept of “chain thinking,” a reasoning methodology developed in 2022 by Google DeepMind, which demonstrated how logically splitting a problem significantly improves the correctness and consistency of the answers provided by language models. IBM is taking this innovation to the next level by applying reinforcement learning techniques directly to the Granite-3.1-8b-instruct model, without resorting to distillations based on larger models, as is done with DeepSeek-R1 and its derivatives. This allows the overall performance of the model to be preserved without compromising reliability and security.

Comparison with other models, such as Llama and Qwen, highlights that Granite maintains high performance on reference benchmarks such as ArenaHard and AlpacaEval, ensuring an optimal balance between advanced reasoning capabilities and operational speed. While approaches based on DeepSeek show excellence in specialized fields such as mathematics and programming, they often sacrifice versatility and general security. A crucial aspect of the innovation introduced is the selective activation of reasoning: developers can decide when to enable the function, adapting the depth of analysis to operational needs. This approach addresses a central problem in high-level inference models: computational cost and processing time. Situations in which a rapid response is preferable – such as the simple geographical location of a city – do not require an extensive reasoning process, thus avoiding waste of resources.

Accessibility is another distinctive element of this preview, made available on IBM watsonx.ai and on the official Hugging Face page under an open Apache 2.0 license. IBM invites the developer and research community to test and provide feedback on these emerging capabilities, helping to improve future versions of the Granite family.

A step forward that combines innovation, flexibility and security in a single solution.