Generative AI for business is a cutting-edge technology that enables businesses to harness the power of AI in order to create, train, and improve models for various business tasks. It emphasizes the importance of trust, security, and adaptability in AI, and introduces Watsonx as IBM's integrated data and AI platform. With Watsonx, businesses can seamlessly access, train, validate, and deploy AI models across hybrid cloud architectures.
The video discusses the importance of AI in big business and the need for AI to be trusted, secured, and adaptable. It highlights the capabilities of AI to work across organizations, scale across systems, and show its work. The speaker emphasizes the role of AI value creators who can bring their own data and models to Watsonx, train and improve them, and own the value created by their models. The video also introduces Watsonx as IBM's integrated data and AI platform, consisting of Watsonx.data, Watsonx.ai, and Watsonx.governance. The platform enables seamless data access, training, validation, and deployment of AI models, and is built on Red Hat OpenShift to support hybrid cloud architectures.
>> Welcome to IBM THINK 2023!
>> AI generated art, AI generated songs.
AI, what is that?
It sure is a lot of fun.
But when foundation models are applied to big business, well,
you need to think bigger.
Because AI and business needs to be held to a higher standard.
Built to be trusted, secured, and adaptable.
This isn't simple automation that is only
trained to do one thing.
This is AI that is built and focused to
work across your organization.
This isn't committing to a single system.
This is hybrid ready AI that can scale across your systems.
This isn't wondering where an answer came from.
This is AI that can show its work.
When you build AI into the core of your business, you
can go so much further.
This is more than AI.
This is AI for business.
>> Please welcome Senior Vice President and
Director of Research, IBM, Dr. Dario Gil.
>> DARIO GIL: Hello.
Welcome, welcome to the last session of THINK.
And I understand some of you even had a drink.
So, I hope you've enjoyed the last two days with us.
And what an incredible year it has been for AI.
You can really feel the change that is happening all around us.
And there's just no denying that the pace of this technology
continues to be exhilarating and that its implications are
now so clear for all to see around the globe.
I am just fascinated by AI.
And as a technologist, this level of excitement really
comes about only maybe once or twice every decade.
And I am just thrilled to see all the possibilities that this
technology is going to enable.
Because it's really going to impact every industry.
From customer care to transforming data centers
and logistics to medicine, to manufacturing, to
energy, to the automotive industry, to aerospace,
communications, you name it.
It's really going to impact every one of our businesses
and really touch every aspect of our lives.
So, it's really exciting.
And while sometimes the pace of this technology can feel,
you know, daunting and scary, the opportunities to harness
foundation models and generative AI with proper governance, the
opportunities are immense.
The emergence of foundation models and generative AI is
really a defining moment.
And we need to recognize its importance, we need
to capture the moment.
And my advice is, don't be just an AI user.
Be an AI value creator.
Just think about it, as an AI user, you are limited
to just prompting someone else's AI model.
It's not your model, you have no control over the
model or the data.
Just think carefully about whether that's the world
you want to live in.
As an AI value creator, on the other hand, you have
multiple entry points.
You can bring your own data and AI models to Watsonx or
choose from a library of tools and technologies.
You can train or influence training, if you want.
You can tune, you can have transparency and control
over the governing data and AI models.
You can prompt it, too.
Instead of only one model, you will have a family of models.
And through this creative process you can improve them
and you can make them your own, your own models.
Foundation models that are trained with your data will
become your most valuable asset.
And as a value creator, you will own that and all the
value that they will create for your business.
So, don't outsource that.
You can simply control your destiny with foundation models.
So, let me show you how we become, and allow you to become
a value creator with Watsonx.
Watsonx is our new integrated data and AI platform.
It consists of three primary parts: First, Watsonx.data it is
our massive curated data repository that is ready to be
tapped to train and fine-tune models, with state-of-the-art
data management system.
Watsonx.ai, this is an enterprise studio to train,
validate, tune, and deploy traditional machine learning and
foundation models that provide generative capabilities.
And Watsonx.governance, this is a powerful set of
tools to ensure your AI is executing responsibly.
Watsonx.data, Watsonx.ai, Watsonx.governance, they
work together seamlessly throughout the entire
lifecycle of foundation models.
And true to our commitment to hybrid cloud architectures,
Watsonx is built on top of Red Hat OpenShift.
Not only does it provide seamless integration of
Watsonx components, it allows you to access and deploy
your AI workloads in any IT environment, no matter where
they are located.
WatsonX is the AI platform for value creators.
And look, I don't need to tell you that deploying these
technologies is not easy at the enterprise level.
But the platform changes that.
So let's take a look now at how an entire AI workflow end to
end works in the platform.
The lifecycle consists of preparing our data, using it to
train the model, validate the model, tune it, and deploy in
applications and solutions.
So let's start with data preparation.
So say you're a data scientist and want to access the data that
is in a Public Cloud, some that is on prem, some that may be in
another external database, or in a Public Cloud, a second one, or
anywhere else outside your hybrid cloud platform.
So you access the platform from your laptop and
It establishes the necessary connections between the data
sources so you can access the data easily.
We have been building our IBM data pile combining raw data
collected from public sources with IBM propriety data.
We are bringing data from different domains, the internet,
code, academic sources, enterprise, and more.
We have used Watsonx.data to collect petabytes of data across
dozens of domains to produce trillions of tokens that we can
use to train foundation models.
And besides the raw data and our proprietary data, we allow
clients to bring their own data to enrich and improve their
purpose-built foundation models.
It is all stored in .data.
With granular metadata that provides traceable governance
for each file or document.
So, now we took this and we move to filter and process the data.
First, we identify the provenance and the idea
of the data.
Then, we need to categorize it, we classify it, for
example, in a pile for different languages, let's say English,
Spanish, German, and so on.
A pile of code data that we then separate by programming
language, Java, Ansible, COBOL, and so on.
And any other category that we have.
Now we filter it, we do analytics and get
rid of duplicated data.
Now identify hate, abuse, and profanity in the data,
and we remove it.
We filter it for private information, licensing
constraints, and data quality.
By annotating, we allow data scientists to
directly determine the right thresholds for their filtering.
Having done all of that, the pile is now ready
for the next step.
We version and tag the data.
Each dataset, after being filtered and preprocessed,
receives a data card.
The data card has the name and the version of the pile,
specifies its content, and filters that have
been applied to it.
And any other relevant content to make it easy to manage and
track different choices that of the work and the right subsets
of the data that we have used to develop the foundation models.
Now, we can have multiple data piles.
They coexist in .data and access the different versions of data
for different purpose is managed seamlessly.
So, we are now ready to take the pile and start
training our model.
This is step 2 in our AI workflow.
So, we move from .data to .ai and start by picking a model
architecture from the five families that IBM provides.
These are bed rocks of models, and they range from
encoder only, encoder- decoder, decoder only, and
other novel architectures.
Let's pick the encoder-decoder sandstone to train the model and
pick a target data pile version from the piles that is .data.
.ai allows training with computing resources across
the hybrid cloud.
In this case it runs on IBM Vela.
Vela is the first of a kind cloud native AI super computer
that we built last year.
It gives you bare metal performance in the cloud with a
virtualization overhead that is less than 5%.
And we are making it available as a service.
Watsonx.ai auto scales the resources for the
training being done.
And the first thing that we need to do is to tokenize the
data according to the requirements of the model.
So, we first query the data using the version ID for
the pile we want to use.
That materializes a copy of the dataset on
Vela for tokenization.
What this means is that, for example, we were building a
large language model, the sentences in the data are
broken into tokens.
And this process can create trillions of them.
And we use the tokens to train the model.
Now, training is a very complex and time consuming task.
It can require dozens, hundreds, even thousands of
GPUs and can take days, weeks, and even months.
Training in Watsonx.ai takes advantage of the best
open-source technology out there to simplify the user experience.
Built on code flare, using PyTorch and Ray, it
also integrates Hugging Face to bring you a rich
variety of open formats.
Once training is done, the model is ready for validation.
So for each model we train, we run an extensive set of
benchmarks to evaluate the model quality across
a wide range of metrics.
Once the model passes all the thresholds across the
benchmarks, it is packaged and marked as ready for use.
For each model, we create a model card that lists all
the details of the model.
We will have many different models, trained
on different piles, with different target goals.
Next, we go to Watsonx.governance to combine
the data card that has the detailed provenance information
for the data pile that was used for training, with the model
card that has the detailed information on how the model
was trained and validated.
Together, they form a fact sheet.
This fact sheet is cataloged in .governance and all the other
fact sheets for all the models that we have available for use.
Now let's go on to tune the model that we just created, and
what we mean by that is to adapt it to new downstream tasks,
which is the basis for the large productivity gains that is
afforded by foundation models.
So, say, in this case, you are a different person, and you are
the application developer.
So, you can access Watsonx.ai and start by picking a model
from the catalog to work with.
We have a family of IBM models specialized
for different domains.
But we also have a rich set of open models, because we
believe in the creativity of the global AI community
and in the diversity of models it offers, and we
want to bring that to you.
In this case, we pick sandstone.3b from the IBM
language models, which is the model that we just trained.
We set up the options for tuning, the tuning approach.
We pick summarization as an example, as the
base model to use.
Now, we can access and use business proprietary data to
tune the base model and for the task that we choose,
whether that business data is located in anywhere in the
hybrid cloud platform.
So, now we send prompts and tuning data, and that's used
to tune the model in .ai.
You get the outcome of the prompt on the model.
This process happens back and forth, back and forth
many times, and in the end, you end up with a set of
ideal prompts to use.
The model is now specialized and ready to deploy.
This is the final step in our AI workflow.
The application where you want to use the foundation model can
live in the public cloud, it can live on-prem or on the edge.
And you can really deploy and run
foundation models efficiently wherever you need them.
And the deployed model can be used in
many different applications.
So, for example, we've embedded foundation models
in Watson Assistant.
For text generation in Assistant, you describe the
topic that you want the assistant to handle, and it
generates the corresponding conversational flow.
We have an inference stack to scale the serving of
the model in applications.
It consists of state-of- the-art technology that has
been field tested for scalable model serving.
This is how Watsonx allows us to go from data to a model that is
trusted, governed, deployed and ready to serve, and how
we can scale that model to different applications.
Once models are deployed, we continuously monitor them
and update them in both .data and in .ai.
We call this constant process our data and model factory.
At Watsonx.governance monitors the models, if there's any
change that may impact how the model can be used or performs,
be driven because we have new data that can
be leveraged or there's a change in some regulation
or law or data licensing.
Any change detected by the .governance process guides and
process the update to both the data and the model.
The idea of the model factory is that is central to solid
and proper governance of AI.
Now, all of these updates need to happen without
disrupting the underlying applications that are leveraging
the foundation models.
And this data and model factory is in production today.
We have already produced over 20 models across modalities like
language, code, geospatial and chemistry, and spanning
different sizes of models from hundreds of millions to
billions of parameters.
We have infused these foundation models into IBM
products, Red Hat products, and our partners' products.
At IBM, over 12 foundation models are powering our IBM NLP
library, which is used in over 15 IBM products and
is available to ISVs.
Granite models train over code are part of IBM Watson
Code Assistant, which has been applied in the Red Hat
Ansible Automation Platform.
And as you heard earlier in this event, SAP has partnered with us
and is infusing foundation models into their solutions.
So, Watsonx is really ready for you to create value with AI.
Now, to maximize what you can do and the innovations at your
disposal, we believe that you should bet on community.
Because, the truth is, one model will not rule them all.
And with the innovations and models that it develops,
the open community is super charging the value that you
will be able to create.
To be true to our belief in the diversity and the creativity of
the open AI community, we are proud to announce our new
partnership with Hugging Face.
So let's invite to the stage cofounder and CEO of
Hugging Face, Clem Delangue.
>> CLEM DELANGUE: Hey, Dario.
>> DARIO GIL: Clem.
>> CLEM DELANGUE: Thanks for having me.
>> DARIO GIL: First of all, welcome to IBM THINK.
We are just delighted to have you here.
So let's begin by, tell us a little bit about yourself
and how and when you got interested in AI and how did
Hugging Face get started.
>> CLEM DELANGUE: Yeah, thanks so much for having me.
I, actually, started in AI almost 15 years ago.
I look at the room at the time we couldn't have filled it.
Maybe it would have been one person, two persons
in the room at most.
As a matter of fact, we weren't even calling it AI
at the time, we were calling it computer vision.
I was working at French company – I am French, as you can hear
from my accents – and we were doing computer vision
on device, on mobile.
The company went on to get acquired by Google after.
But I never lost my passion and excitement for AI.
So, seven years ago, with my cofounders, Jillian Thomas, we
gathered around this passion for AI and started Hugging Face,
right, what you see on my T-shirt, basically.
We started with something completely different.
We worked on conversational AI for three years and as it
sometimes happens for startups, the underlying platform and
technology ended up more useful than the end product.
When we started to release part of it on GitHub, we started to
see open-source contributors joining us, we started to see
scientist sharing models in the platform, leading to what
Hugging Face is today.
>> DARIO GIL: So I mentioned the power and the creativity of the
open community creating in AI.
Just share with us some statistic, how big is it?
How much energy is there in that community and how much should we
expect in the creativity available to all of us?
>> CLEM DELANGUE: Yeah, the energy in open-source AI
is insane these days.
Just a few weeks ago I was in San Francisco.
I tweeted that I would be around and that we could do some
sort of a small get-together for open-source AI people.
We thought we would get maybe a few dozen, few hundred people.
And the more the days came, the more people ended up joining.
We had to change locations three times to something at
the end almost as big as that, we had 5000 people.
People started calling it the Woodstock of AI,
so that's just an example.
We are competing with that, the Woodstock of AI.
Just proof of how vibrant the open-source AI community is.
We think the same thing on Hugging Face, right?
Since we started on the platform four years ago, we grew to now
having over 15,000 companies using the platform including
very large companies like Google, like Meta, like
Bloomberg, all the way down to smaller companies like
Grammarly, for example.
And collectively they have shared over 250,000 open
models on the platform, 50,000 datasets, and over
100,000 open demos.
Just last week 4000 new models have been
shared on the platform.
So, that shows you kind of like the magnitude and energy in
open-source AI community.
>> DARIO GIL: Just think about that, 4000 models in one week.
So, one of the myth busting things that we were chatting
about is that the element of one model will not
rule them all, right?
There's going to be a huge amount of innovation that is
happening from so many sources.
So, perhaps, you could share with us, what are some examples
of innovation that you see?
We have seen scale.
But what are some examples that really caught your
eye or you think were particularly powerful?
>> CLEM DELANGUE: Yeah, I mean, it's interesting because since
the release of ChatGPT, right, and some people have
said, okay, ChatGPT is a model to rule them all.
100,000 new models have been added on Hugging Face, right?
And, obviously, companies, they don't train models just
to train models, right?
They would prefer not to do it because it costs
money to train models.
But the truth is, if you look at how AI is built, when you can
build smaller, more specialized, customized models for your use
cases, they end up being cheaper, they end up being more
efficient, and they end up being better for your use case, right?
Just the same way every single technology company learned how
to write code, right, and to have a different code base than
their competitors or than companies in other fields.
We are seeing the same thing for AI, right?
Every single company needs to train their own models, optimize
their own models, learn how to run these models at scale.
Every single company needs to build their own ChatGPT because
if they don't, they won't be able to differentiate, they
won't be able to create the unique technology value that
they have been building for their customers, and they lose
control, right, if they start outsourcing it.
That's what we are seeing on Hugging Face and in the
ecosystem as a whole.
>> DARIO GIL: It's back to the philosophy of don't be a prompt
tuner user, right, be a value creator with all of this.
So let's talk about our partnership for a minute.
Why are you excited about bringing the power of all of
this community into Watsonx, in the context now of an
enterprise, you know, need and meeting the needs of our clients
that are here listening?
>> CLEM DELANGUE: Yeah, obviously, Hugging Face and IBM
share a lot of the same DNA, right, around open-source,
open platform, kind of, like, providing extensible
tools for companies.
For me, one of the most iconic collaboration partnership of the
last decade is IBM plus Red Hat and hopefully we are just at the
beginning of it, but with this collaboration, we can
do the same thing for AI.
I think with this integration between Watsonx and Hugging
Face, you kind of like get the best of both worlds in the
sense that you get the cutting edge and the community and
the numbers of models, datasets, apps of the
Hugging Face ecosystem, and you get the security and
supports of IBM, right?
For example, you mentioned, we mentioned all the models.
The IBM consultants can help you to pick the right models for you
at the time that is going to make sense for your company.
So, you really get, kind of, like, the perfect mix to get to
what we were saying, meaning every one of you being able to
build your own internal ChatGPT.
>> DARIO GIL: So, tell us, this is just great.
I am just delighted about those opportunities.
So tell us a little bit about what's next for Hugging Face
when you look over the next year or so, what
excites you the most?
>> CLEM DELANGUE: Many, many exciting things for us.
We have seen a lot of adoption, a lot of companies using us for
text, for ODO, for image.
And now we are starting to see that expand to other domains,
for example, we are seeing a lot of video right now.
We are seeing a lot of recommender systems; we are
seeing a lot of time series.
We are starting a lot of bioG chemistry.
We are excited about it, we think ultimately AI is the new
default to build all features, all workflows, all products.
It's kind of like new default to build all tech.
So we are excited for this expansion to other domains.
Also, we are seeing a lot of work around chaining different
models and, in fact, at Hugging Face we released today a
transformer agents which is a way to chain different models to
build more complex systems that are achieving kind of
like better capabilities.
These are some of the things that we are the
most excited about.
>> DARIO GIL: So a lot there so thank you, Clem.
Thank you so much.
>> CLEM DELANGUE: Thank you so much.
>> DARIO GIL: Thank you.
So, while you saw how the platform works to enable the
foundation model creation workflow end to end.
And we talked about data, we talked about model
architectures, the computing infrastructure, the models
themselves, the importance of the open community.
So, now let me show you how to use and how you
would experience Watsonx.
And we are going to go inside the studio, inside Watsonx.ai.
And from the landing page you can choose two prompt models to
fine-tune models or deploy and manage your deployed models.
So here's an example of how you can use the prompt lab to
do a summarization task.
You give the model the text as a prompt, and the model
summarizes it for you.
In the case of a customer care interaction, it gives you the
customer problem and the resolution according to the
transcript of the interaction.
In the tuning studio, as we saw before, you can set the
parameters for the type of tuning that you want to do
and the base model and you can add your data.
The studio gives you detailed stats of the tuning process and
allows you to deploy the tune model in your application.
It's that simple.
We took the complexity of the process away so you only
need to worry about creating value for your business.
And here are some of our current AI value creators.
SAP will use IBM Watson capabilities to power its
digital assistant in the recipe solutions.
You have been hearing about Red Hat, how it's embedding IBM
Watson Code Assistant into the Ansible Automation Platform,
BBVA is bringing their enterprise data to use with
their own foundation model for natural language.
Moderna is applying IBM's foundation models
to help predict potential MRNA medicines.
NASA is using our language models together with US spatial
models we have created together to improve our
scientific understanding and response to earth and
climate related issues.
And WiX is using foundation models to gain novel insights
for customer care as they meet the needs of their customers.
So, what I encourage you is to join them and embrace the age
of value creation with AI.
A year ago, I stood on a stage just like this, closing THINK.
And I shared with all of the attendees that what was next in
AI was foundation models.
And maybe at the time it seemed a little bit abstract and, you
know sort of, like, this intellectual disposition about
where things were going.
But, boy, what a year it has been.
And it has been a big year for AI at IBM.
So, as we close our event this year, let me remind of you of
all of the things we have created and announced.
We have announced Watsonx, a comprehensive platform that
allows you to create and governor AI in real time so that
you can move with urgency and capture this moment.
We announced a set, a family of foundation models, including
IBM models, open community models, and how you can even
create your own models.
We announced our data model factory, using petabytes of data
across multiple domains to create trillions of tokens to
create our family of foundation models and show how the factory
continuously updates them when conditions change and brings a
regular cadence of models to ensure proper governance.
We told you about products where we have infused our
foundation models over 15 of them, including digital labor,
Red Hat products like Ansible Automation Platform, our partner
products like ACP solutions.
We announced important collaborations to advance AI and
bring it to the enterprise, Hugging Face and long-standing
collaborations and initiatives like PyTorch and Ray.
We showed you some of the organizations that have become
AI value creators with us.
We are bringing IBM Vela or cloud native AI super computer
to train foundation models with bare metal performance
while giving us the flexibility of the cloud.
And we announced that we are making it
available as a service.
Last year we launched the Telum NC16.
It's an engineering marvel and IBM's first processor to
have on chip accelerator for AI inferencing.
It can process 300 billion inference per day
with one millisecond latencies.
This means now you can infuse AI into every transaction
in Z16 for applications like fraud detection and
others in real time.
Using the same core architecture as Telum, we built the IBM
Research AIU, which is optimized to give superior performance for
foundation models and enable with Rat Hat software stack.
And at IBM Research we are incubating powerful
AIU systems designed and optimized for enterprise, AI
inference, and tuning.
So a truly fantastic year and this is just the start of all
the amazing things that we are building and developing for you
and that we will be sharing with you in the coming years.
So, today more than ever before, it's important to have a
business strategy in AI.
And in closing, as you think about how to harness foundation
models for your business, let me offer you some tips to consider.
First, act with urgency.
This is a transformative moment in technology, be
bold and capture the moment.
Second, be a value creator, build foundation models on your
data and under your control.
They will become your most valuable asset.
Don't outsource that and don't reduce your AI
strategy to an API call.
Third, bet on community.
Bet on the energy and the ingenuity of the
open AI community.
One model, I guarantee you, will not rule them all.
Run everywhere efficiently, optimize for performance,
latency, and cost by building with open hybrid technologies.
And finally, be responsible.
I can't stress this enough.
Everything I have mentioned is useless unless you build
responsibly, transparently, and put governance into the
heart of your AI lifecycle.
Continuously governor the data you use and the AI you deploy.
And co-create with trusted partners, trust is your
ultimate license to operate.
If you map your AI business strategy against these
recommendations, you will be in a prime position to do amazing
things with foundation models and generative AI.
We have built Watsonx so that you can do just that.
And I hope you join us, because we cannot wait to get started
on this journey with you.
In this year's conference, the focus is on the incredible potential of AI in businesses. AI has become more than just a source of entertainment; it is now being applied to transform industries and touch every aspect of our lives.
AI has the power to revolutionize various industries, including customer care, data centers, logistics, medicine, manufacturing, energy, automotive, aerospace, communications, and more. The possibilities are immense, and its impact is undeniable.
As an AI user, you have limited control over the models and data. But as an AI value creator, you have multiple entry points and the ability to bring your own data and AI models to the table. You can train, tune, prompt, and improve the models to make them your own. With foundation models trained on your data, you own a valuable asset that can create immense value for your business.
IBM's new integrated data and AI platform, Watsonx, empowers you to become an AI value creator. It consists of three primary parts:
These components work seamlessly together throughout the entire lifecycle of foundation models. Watsonx is built on Red Hat OpenShift, providing integration and accessibility across different IT environments.
The AI workflow within Watsonx enables you to prepare, train, validate, tune, and deploy your models. You can access and process data from various sources, filter and preprocess it, version and tag it, and use it to train your models. Watsonx.ai leverages computing resources across the hybrid cloud, such as IBM Vela, to tokenize and train the models efficiently. Once trained, the models undergo extensive validation and benchmarking before being marked as ready for use.
With Watsonx, you have the power to control your destiny in the AI realm. By becoming an AI value creator, you can harness the immense potential of AI for your business and make a meaningful impact. Don't just be an AI user; be an AI value creator with Watsonx.
Five effective strategies for generating instant sales without ads are discussed in the text. These include creating an affiliate program, leaving comments and responses on the web, writing controversial blog posts, running joint webinars, and publishing videos on YouTube with keyword optimization. These strategies have been proven to drive customer engagement and increase revenue.
Unified Communications as a Service (UCaaS) offers streamlined communication tools for businesses, improving productivity and user experience. UCaaS platforms integrate voice, video, messaging, and collaboration features, making it easier to communicate and work together remotely. Implementing UCaaS requires careful consideration of connectivity, service level agreements, and interoperability. However, the benefits include cost savings, scalability, and future-proofing business communications.
Using AI, the author analyzes their webpage to determine how well it aligns with their audience's information needs for SEO and conversion optimization. The AI then creates a table showing the page's ability to inform and persuade, as well as its relevance to key phrases. The AI also generates an outline for a major revision to the page, aiming to achieve top scores in all categories. The author can further train the AI and compare drafts before and after, ultimately evaluating the success of the page through website metrics like traffic, engagement, and conversion rate.
Embracing AI and leveraging first-party data can improve targeted advertising and customer engagement in digital marketing.
We appreciate your recent sign up for a LiveAgent.
A message will be sent to your email address containing login details, right after your account is installed.
We’re available on multiple dates