-
Generative AI
Generative AI is here:
Are you ready?Intel® Innovation Built-in
-
-
-
Where to start
A new Generative AI age
Generative AI (GenAI) is being leveraged on a massive scale by organizations and individuals, causing a significant societal impact. Consumer-grade AI, like ChatGPT and DALL-E, has captured everyone's imagination with its ability to generate content. Yet, GenAI's impact on organizations promises even more value, including heightened productivity, cost reduction, and a transformation in how we work.
-
GenAI will transform organizations
GenAI brings rewards, but it also comes with new challenges and risks. As organizations embark on their GenAI journey, they cannot risk customer trust and the high value of their data for the reward of being first to the finish line.
-
-
-
Achieve Generative AI success
Landing on the right use cases is crucial. Business and IT leaders should prioritize use cases based on these criteria:
-
-
-
The vital role of your data in Generative AI
Data and risk go hand-in-hand. Data will drive your GenAI projects forward, but you also need to assess the potential risks of hosting GenAI models in public clouds, including: intellectual property loss, data leakage, privacy issues, compliance violations, credibility and integrity loss, bias, and IP infringement.
-
-
Managing risk and enhancing value
As you begin your journey, it is essential to align investments in technology and training to boost operational maturity, reduce risk, enhance control, and maximize value for your organization. With enterprise-ready GenAI, you gain control over who can access your data.
-
-
Enterprise-ready GenAI Immature GenAI Operational Maturity Risk Graph showing that Risk (y-axis) decreases over time as you reach Operational Maturity (x-axis) with your data handling. The start of the slope of the graph indicates Immature GenAI with High Risk and Low Operational Maturity. The downward trend on the graph indicates that Enterprise-ready AI decreases Risk and represents High Operational Maturity.
-
-
-
-
-
USE AI WITH YOUR DATA
Keep Generative AI models close to your data
-
-
Understanding the risks and benefits of different deployment options is crucial in determining your organization’s optimal workload placement for GenAI. When it comes to bringing AI to your data, deploying private instances of GenAI large language models (LLMs), such as Llama 2 or Falcon, offer advantages in speed and deployment, but they may also involve higher costs and other drawbacks. Either way, in-house GenAI will likely provide the most value for your early efforts.
In terms of workload placement, GenAI is no different than any other workload. To get the best outcomes, put it in the environment that makes the most sense based on your business requirements and technical needs.
The diagram below conveys concepts and frameworks that come into play when determining GenAI workload placement.
-
-
-
5 reasons why you should bring AI to your data
-
-
-
-
Where data resides Cost Faster time to value Accuracy and customization General purpose use cases Secure access Public Cloud Private Cloud A chart representing six factors to consider when choosing between Private Cloud or Public Cloud for GenAI workload placement. ‘Where data resides’ and ‘Secure access’ lean greatly toward Private Cloud. ‘Cost’ leans moderately toward Private Cloud and ‘Accuracy and customization’ leans slightly to Private Cloud. ‘Faster time to value’ leans slightly to Public Cloud. And ‘General purpose use cases’ leans greatly to Public Cloud.
-
-
Data Management for Generative AI
Most organizations are taking a two-pronged approach to their GenAI strategy. They're experimenting with tactical deployments to learn and avoid falling behind, while also developing a long-term strategy to accommodate the many use cases that will emerge over time. This approach requires a two-tiered data management strategy.
DATA PREPARATIONDATA ENGINEERING -
-
-
Data Preparation
-
Short-term: Data Preparation
Data preparation includes identifying data sets and defining data requirements followed by cleansing, labeling, and anonymizing the data, then normalizing it across data sources. It also requires building data pipelines to integrate the data into a model.
-
-
-
Data Engineering
-
Long-term: Data Engineering
Organizations need a well-structured data repository, such as a data lake or data lakehouse, to integrate their data with GenAI models. Consider building the data lake iteratively to progressively expand the capabilities of the GenAI data repository while the team enhances their data management and GenAI skills.
-
-
-
-
-
-
“This collaboration [with Dell Technologies] will empower companies to build their own AI systems leveraging the incredible innovation of the open source community while benefiting from the security, compliance, and performance of Dell systems.”
-
-
-
-
-
-
RIGHT-SIZE AI
Define infrastructure and right-size AI
-
Your unique data enables you to utilize domain- and enterprise-specific use cases, creating industry value through tasks or functions for which you have exclusive ownership of the data. Different types of GenAI have corresponding entry points and investments that are necessary to ensure success. LLMs trained on vast amounts of text are like encyclopedias, helpful for general use, but may not be suitable for answering specific questions regarding your organizational data.
-
-
-
Your data greatly improves the efficiency and value of GenAI
-
Limited functionality Smallest/most cost effective Domain expertise Most accurate Smaller size More accurate More hallucinations Costly & energy intensive Breadth of use cases General purpose Enterprise specific Domain specific Large language models Value Graphic showing relative amounts of data required for 3 types of AI models, as well as their business value. Large Language Models, or LLMs, are for general purpose use cases and use the most data. They can be costly and energy-intensive and they are more prone to hallucinations. Domain-specific AI uses less, but more specific data. It has limited functionality, but is more relevant to your business and has more value. Enterprise-specific AI uses still less data, but is the most specific and accurate, and offers the greatest value to your business.
-
-
-
AI deployment patterns
The AI model you choose will depend on your organization’s level of data science readiness, deployment patterns, and the implications of each.
-
Pre-trained model
Referred to as 'prompt engineering,' this approach involves posing a question to a pre-trained model and receiving a result.
Example: ChatGPTModel augmentation
Enhance your GenAI model by adding your data to provide additional context for its answers, such as inferencing, which includes use cases like Retrieval-augmented generation (RAG).
Fine-tuning models
This involves adjusting model weighting and incorporating your data. While it leads to improved results, it also demands more effort during setup.
Model training
It includes building a specific model and training it with a data set. This typically requires the most work and resources and is often reserved for solving complex problems.
Effort Cost Value &
differentiationData integration Infrastructure Client – server Client – server GPU optimized Large GPU deployment Skills IT Ops Developer Data scientist(s) Data scientist(s) Simplifed deployment reference Validated Design Reference Design Validated Design Reference Design
-
Choose the right infrastructure for your model
The infrastructure supporting your GenAI deployment depends largely on computational requirements, influenced by model type, model size, and number of users. Additional considerations include the necessary storage capacity for data used during deployment, training, and model refinement.
-
GENERAL COMPUTE (CPU Oriented) AI OPTIMIZED (GPU Intensive) Dell Laptops Precision Workstations PowerEdge Rack Servers PowerEdge XE Servers ObjectScale Storage PowerScale Storage PowerFlex Storage PowerSwitch Z Series Switches Model training Fine-tuning model Model augmentation Pre-trained model Millions of parameters Billions of parameters Trillions of parameters Low number of users High number of users Chart representing three GenAI requirements and mapping them to the appropriate Dell hardware solutions. The hardware solutions range in power from General Compute options, which are CPU-oriented, up to AI-optimized options, which are GPU-intensive. The specific options start with Dell Laptops on the General Compute end, and progress through Precision Workstations and PowerEdge Servers, ending with PowerEdge XE Servers on the AI-optimized solutions end. Note that Dell storage and networking hardware can be used across the entire range. Three GenAI infrastructure attributes are mapped to these solutions in a progression that requires more and more processing power. Complexity of Model can range from using Pre-trained Models and Augmenting or Fine-tuning Models to training new Models. Number of Parameters can range from millions to billions, all the way up to trillions.
-
-
-
-
-
Accelerate your ai journey
Start with an early win
Retrieval-augmented generation (RAG) is an ideal starting point for many organizations as it augments a GenAI model with your own data without the process of retraining it. Explore the setup of RAG use cases that can be applied to enhance your business and data.
-
RAG Use Case
Apply RAG to a custom PDF dataset
-
DELL VALIDATED DESIGN FOR RAG
Deploy a digital assistant on Dell APEX Cloud Platform for Red Hat OpenShift
-
-