Granite 4 Hybrid Models: Redefining AI with the Dell AI Factory

Discover how Dell & IBM's Granite 4.0 redefines AI with hybrid efficiency, long-context power, and enterprise-ready innovation!

At Dell Technologies, we believe technology should empower every organization and drive human progress. That’s why we’re announcing our latest collaboration with IBM, bringing the Granite 4.0 family of models to the Dell Enterprise Hub on Hugging Face.

This matters because enterprises are increasingly looking to run “open-source” models on-premises, and the Dell Enterprise Hub paired with the Dell AI Factory offers users a complete, validated path to secure, enterprise-grade AI deployment.

What’s new? The future is hybrid: Mamba2 meets transformer

IBM’s Granite 4.0 leverages a unique model architecture combining Mamba and transformer components to deliver the best of both worlds. Mamba’s efficiency handles the heavy lifting of understanding global context across long documents, while transformer blocks provide the finishing touch, refining local details with high precision.

This powerful combination delivers the long-context capability, speed and memory efficiency of Mamba with the proven accuracy of transformers in a resource-conscious design. This along with the Mixture of Experts (MoE) strategy for select Granite 4.0 models further reduces memory and compute footprint, ideal for enterprises.

provided by submitter

Granite 4.0 on the Dell AI Factory: Innovation simplified

Advanced models are only as good as their accessibility. By offering Granite 4.0 through the Dell Enterprise Hub on our validated, full-stack Dell infrastructure, via the Dell AI Factory, we’re making cutting-edge AI enterprise-ready.

This collaboration has tangible benefits:

    • Radical Efficiency: Based on this benchmark from IBM, comparing Granite 4.0 H Tiny 7B and Granite 3.3 8B  – the memory efficiency is radical at higher context length and higher concurrency.
    • Unlock Long-Context Power: There is a phenomenon called Context Rot, where transformer-based models drastically lose their accuracy at longer context. Because Mamba-2 models perform better at high context in terms of accuracy, this hybrid model architecture removes model-based context limitations.
      • IBM Granite 4.0 models also implemented NoPE (No Positional Encoding), a concept in transformer models where explicit positional encodings are removed. Despite the absence of positional encodings, these models can still learn and infer positional information through other mechanisms within the network allowing it to perform better at long context.
    • Drive Open Innovation: The open Apache 2.0 license empowers developers to freely customize, fine-tune and adapt the models for specific business needs, fostering a culture of open and collaborative enterprise AI.

Meet the Granite 4.0 family

The Dell Enterprise Hub provides access to a portfolio of Granite 4.0 models, each tailored to empower different enterprise applications:

    • Granite 4.0 H Small: With 32 billion total parameters (9 billion active), this hybrid MoE model is a powerhouse for enterprise workloads that require deep contextual understanding of long and complex documents.
    • Granite 4.0 H Tiny: An efficient hybrid MoE model with 7 billion total parameters (1 billion active), optimized for edge deployments and local experimentation. It brings powerful AI to where your data lives.
    • Granite 4.0 H Micro: A compact 3-billion-parameter traditional dense model as an alternative where Mamba-2 support is not yet optimized.

Real-world impact, powered by Dell and IBM

The unique architecture of Granite 4.0 unlocks a new world of possibilities for your business.

    • Deeper Document Insights: By maintaining accuracy at higher context the models go beyond simple summaries. Analyze lengthy legal contracts, complex financial reports, or dense technical manuals to uncover critical insights with unprecedented speed.
    • Smarter RAG Systems: Maintaining accuracy at longer context is critical to Retrieval-Augmented Generation (RAG) based systems. By pairing with Granite Docling, users can build more grounded, efficient, and responsive systems.
    • AI at the Edge: With models like Granite 4.0 Micro, users can deploy real-time AI directly on devices, independent of the cloud. This is a game-changer for industries like manufacturing, retail and logistics, where immediate, onsite intelligence is critical.

Dell and Hugging Face: Simplifying Enterprise AI

Through the Dell Enterprise Hub on Hugging Face, together we are making it easier than ever to deploy, manage and scale AI with confidence.

The Dell Enterprise Hub gives businesses access to pre-validated models optimized for Dell’s AI-ready infrastructure, powered by the latest NVIDIA GPUs. Enterprises can even use Dell Pro AI Studio to manage, deploy, and use Granite 4.0 H Small and Tiny on Dell Pro Max PCs via the Hub, making it easy to use Granite throughout your AI solutions.

Our comprehensive suite of tools, including the Dell AI CLI/Python SDK, Application Catalog and Model Catalog, removes friction from the development process. This complete, end-to-end solution—from hardware to software and support, accelerates your journey from concept to production.

By welcoming innovative models like IBM Granite 4.0 into this trusted ecosystem, we are empowering organizations like yours to build the future of AI on your own terms. Together, we can turn technological potential into human progress.

Learn more about how Dell Pro AI Studio is unlocking powerful on-device AI to your deskside with IBM Granite models. 

Meet us at IBM TechXchange in Orlando, October 6-9, for a live Hybrid AI demo showcasing Granite 4.0 models in action and see how Dell AI Factory can accelerate your journey.

About the Author: Ihab Tarazi

Ihab Tarazi is the Chief Technology Officer and Senior Vice President at Dell Technologies, Infrastructure Solutions Group Core where he is responsible for technology strategy and next generation architecture for Compute, Storage and Network platforms including Edge, 5G and AI/HPC, as well as engagements with technology vendors and cloud ecosystems. Prior to this role, Mr. Tarazi served as Chief Technology Officer at Packet Host, Inc. (currently Equinix Metal), a leading bare metal cloud company built for developers, where he was responsible for the company's Product, Engineering and Technology. Prior to joining Packet, Mr. Tarazi held a number of positions in the telecom and infrastructure space, including Chief Technology Officer at data center and colocation provider Equinix, where he was responsible for developing the Equinix Cloud Exchange and strategic engagements with leading platforms in the Cloud, Security, Storage, Edge and IoT. Mr. Tarazi previously served as VP Engineering and Technology at Verizon, where he led a global team for the Engineering and Product Development of Global Enterprise Services including MPLS VPN, IP, Ethernet, Security and Managed Services. He has served as a member of the board of directors for NeoPhotonics Corp. since October 2015.  He is also a member of the board for Telecom Infrastructure Project (TIP). Mr. Tarazi holds a B.S. degree in Electrical Engineering and Telecommunications from the University of Maryland, and an M.S. in Telecommunications Management from Southern Methodist University.