Learning Library

← Back to Library

Choosing Enterprise LLMs: IBM Granite

Key Points

  • Enterprise‑grade foundation models should be evaluated on three core metrics: performance (latency/throughput), cost‑effectiveness (low inference energy and expense), and trustworthiness (low hallucination and clear training‑data provenance).
  • Trust is especially critical because generative AI workloads can consume 4–5× the energy of traditional web searches, so models must balance high performance with minimal inference cost while offering transparent, auditable training data.
  • IBM’s Granite foundation models are positioned to meet all three criteria equally, delivering strong performance, competitive operating costs, and built‑in trust mechanisms such as documented data sources and reduced hallucination risk.
  • The Granite models are open‑source (Apache 2.0) and trained on vetted enterprise‑grade datasets—including 1.8 M scientific papers, all U.S. utility patents (1975‑2023), and public‑domain legal opinions—totaling about 6.5 TB of data, providing unprecedented transparency and relevance for business applications.

Full Transcript

# Choosing Enterprise LLMs: IBM Granite **Source:** [https://www.youtube.com/watch?v=cVDv9apGTXo](https://www.youtube.com/watch?v=cVDv9apGTXo) **Duration:** 00:06:32 ## Summary - Enterprise‑grade foundation models should be evaluated on three core metrics: performance (latency/throughput), cost‑effectiveness (low inference energy and expense), and trustworthiness (low hallucination and clear training‑data provenance). - Trust is especially critical because generative AI workloads can consume 4–5× the energy of traditional web searches, so models must balance high performance with minimal inference cost while offering transparent, auditable training data. - IBM’s Granite foundation models are positioned to meet all three criteria equally, delivering strong performance, competitive operating costs, and built‑in trust mechanisms such as documented data sources and reduced hallucination risk. - The Granite models are open‑source (Apache 2.0) and trained on vetted enterprise‑grade datasets—including 1.8 M scientific papers, all U.S. utility patents (1975‑2023), and public‑domain legal opinions—totaling about 6.5 TB of data, providing unprecedented transparency and relevance for business applications. ## Sections - [00:00:00](https://www.youtube.com/watch?v=cVDv9apGTXo&t=0s) **Evaluating Enterprise Foundation Models** - The speaker outlines three key criteria—performance, cost‑effectiveness, and trustworthiness—for selecting an enterprise‑grade large language model, citing IBM Granite as an example. ## Full Transcript
0:00when it comes to picking a large 0:01language model with sport for Choice 0:04last time I checked there was something 0:06like 0:08700,000 different llms or large language 0:12models on hugging face now I'd like to 0:14cover just a couple of those 0:17specifically the IBM Granite Foundation 0:20models but first let's consider how to 0:23pick an Enterprise grade Foundation 0:25model meaning an nlm suitable for 0:28deployment in an Enterprise setting 0:30something you'd be happy to run your 0:32business with so let's consider that 0:34through three different metrics so the 0:37foundation model it needs to be 0:41performant that's an important metric 0:44but it also needs to be cost effective 0:49and it needs to be trusted those are the 0:53three metrics we're going to consider 0:55and trust it of course because you can't 0:57scale generative AI with models that you 1:00can't trust so take these one by one now 1:04by performance we're talking about 1:07measurements like latency and throughput 1:10is a foundation model able to keep up 1:12with the speed and Enterprise requires 1:15it to operate that then related to that 1:19is cost Effectiveness now according to 1:22the scientific Jour nature a search 1:25that's driven by generative AI will use 1:28something like 4 25 times the amount of 1:31energy that's needed to run a 1:33conventional web search so we need a 1:36foundation model that can deliver the 1:37necessary performance with low 1:40inferencing costs and we need the 1:43foundation model to be trusted and we 1:47can gauge that through metrics like 1:48hallucination scores but also a model 1:50that offers transparency so we know what 1:53data the model was trained on and I 1:56think in many instances models are kind 1:59of SK 2:00a bit like this they're highly 2:04performant but they're expensive to run 2:07at inference time and there's a lack of 2:10transparency on the training data the 2:12model was built with now with the 2:15granite models IBM set out to create 2:17Enterprise grade Foundation models that 2:19apply an equal weight to all three of 2:23these metrics so it looks more like this 2:28so what should you know about the ABM 2:30Granite Foundation models well many of 2:32the models are open source you can find 2:35them on hugging face under the Apache 2:382.0 license that enables broad 2:40commercial usage now these models also 2:43have transparency in training data 2:47meaning we actually know the data 2:49sources that we use to train the models 2:51and that's quite atypical most llms are 2:55uh notoriously vague on how their models 2:58were trained so that's a nice change now 3:00Granite language models are trained on 3:02trusted Enterprise data spanning 3:05academic code legal and finance data 3:08sources as such as well the first 13 3:13billion parameter Granite llm was 3:15trained on about 3:186.5 terabytes of data and that includes 3:231.8 million scientific papers that were 3:27posted on archive it ALS o includes all 3:30us utility patents granted by the 3:35USPTO and that's from 1975 all the way 3:38through to 2023 and it includes the 3:41public domain free 3:45law which are legal opinions from US 3:48federal and state courts essentially the 3:51models have been governed and filtered 3:53to only use Enterprise safe data sources 3:57the granite models have also been 3:59designed to to be performant as well 4:03especially in areas of coding and 4:06language tasks outperforming some models 4:08that are actually twice their size and 4:10smaller models means also they're more 4:14efficient with less compute requirement 4:17and a lower cost of inferencing now I 4:20keep mentioning the granite models 4:22plural so which models are we talking 4:25about so Granite is actually a family of 4:29llm Foundation models spanning multiple 4:32modalities and you can find many of 4:34these on hugging face so let's take a 4:37look at some of them and we'll start 4:39with granite for language now these are 4:44decoder models of different parameter 4:47sizes so that includes a 4:507B open source model and the B here 4:54refers to billions of parameters so 4:56seven billion parameters there's also an 5:008B model that's designed specifically 5:03for Japanese text there's a couple of 5:0713B models and there is a 20 billion 5:12parameter multilingual model that 5:14supports English German Spanish French 5:17and 5:18Portuguese now there's also Granite for 5:23code and that again comes in different 5:26parameter sizes from 3 billion all the 5:30way through to 5:3134 billion parameters and granite for 5:35code is trained 5:37on6 programming languages now there's 5:41also Granite for time series that's a 5:47family of pre-trained models for time 5:50series forecasting these models are 5:52trained on a collection of data sets 5:53spanning a range of business and 5:55Industrial application domains and these 5:57models are optimized to run on pretty 6:00much anything even a laptop and then 6:03finally there is granite for geo spatial 6:08which is a partnership between NASA and 6:11IBM to create a foundation model for 6:13Earth observations using large scale 6:15satellite and remote sensing data so 6:18that's the IBM Granite models models 6:21that are trusted performant and 6:23efficient and that can be applied to a 6:25wide variety of Enterprise use 6:28cases for