Free From Google: Generative AI Learning Path

Illustration by Author

Are you interested in discovering the potential of Generative AI Models and their applications? Luckily, Google Cloud released the Generative AI Learning Path, a great collection of free courses, that start from explaining the basic concepts of Generative AI to more sophisticated tools, like Generative AI Studio to build your customized generative AI models.

This article will explore seven of the available courses, that will allow you to understand the concepts behind the Large Language Models that surround us every day and create new AI solutions. Let’s get started!

Course link: Introduction to Generative AI

This first course is an introduction to Generative AI by Dr Gwendolyn Stripling, an AI Technical Curriculum Developer at Google Cloud. It will make you learn what Generative AI is and how it’s applied. It begins with the basic concepts of data science (AI, Machine Learning, Deep Learning), and how Generative AI is different from these disciplines. Moreover, it explains the key concepts that surround Generative AI with very intuitive illustrations, such as transformers, hallucinations and Large Language Models.

Video duration: 22 minutes

Lecturer: Gwendolyn Stripling

Suggested readings:

Course link: Introduction to Large Language Models

This second course aims to introduce what Language Models are at a high level. In particular, it gives examples of LLM applications, like text classification, question answering and document summarization. In the end, it shows the potential of Google’s Generative AI Development tools to build your applications with no code.

Video duration: 15 minutes

Lecturer: John Ewald

Suggested readings:

Course link: Introduction to Image Generation

This third course focuses on explaining the most important diffusion models, a family of models that generate images. Some of the most promising approaches are Variational Autoencoders, Generative Adversarial Models and Autoregressive Models.

It also shows the use cases that can be categorized into two types: unconditioned generation and conditioned generation. The first includes human face synthesis and super-resolution as applications, while examples of conditioned generation are the generation of images from a text prompt, the image inpainting and the text-guided image-to-image.

Video duration: 9 minutes

Lecturer: Kyle Steckler

Course link: Attention Mechanism

In this short course, you will learn more about the attention mechanism, which is a very important concept behind transformers and Large Language Models. It has enabled improving tasks, such as machine translation, text summarization and question answering.  In particular, it shows how the attention mechanism works to solve the machine translation.

Video duration: 5 minutes

Lecturer: Sanjana Reddy

Course link: Transformer Models and BERT Model

This course covers transformer architecture, which is an underlying concept behind the BERT model. After explaining the transformer, it gives an overview of BERT and how it’s applied to solve different tasks, such as single-sentence classification and question answering.

Differently from the previous courses, the theory is accompanied by a laboratory, which requires prior knowledge of Python and TensorFlow.

Video duration: 22 minutes

Lecturer: Sanjana Reddy

Suggested readings:

Course link: Create Image Captioning Models

This course aims to explain the Image Captioning Models, which are generative models that produce text captions by taking images as input. It exploits an encoder-decoder structure, attention mechanism and transformer to solve the task of predicting a caption for a given image. Like the previous course, there is also a laboratory to put the theory into practice. It is again oriented toward data professionals with prior knowledge in Python and Tensorflow.

Video duration: 29 minutes

Lecturer: Takumi Ohyama

Course link: Introduction to Generative AI Studio

This last course introduces and explores Generative AI Studio. It starts by explaining again what Generative AI is and its use cases, such as code generation, information extraction and virtual assistance. After giving an overview of these core concepts, Google Cloud shows the tools that help on solving Generative AI tasks even without an AI background. One of these tools is Vertex AI, which is a platform that enables to manage the machine learning cycle, from building to the deployment of the machine model.  This end-to-end platform includes two products, Generative AI Studio and Model Garden. The course is focused on explaining Generative AI Studio,  which allows to build easily generative models with no code or low code.

Video duration: 15 minutes

Suggested readings:

I hope that you have found useful this fast overview of the Generative AI course provided by Google Cloud. If you don’t know where to start in understanding the core concepts of Generative AI, this path covers every aspect. In case you have already a machine learning background, there are surely models and use cases you can discover from one of these courses. Do you know other free courses about Generative AI? Drop them in the comments if you have insightful suggestions.

Eugenia Anello is currently a research fellow at the Department of Information Engineering of the University of Padova, Italy. Her research project is focused on Continual Learning combined with Anomaly Detection.

Source link

Leave a Reply

Your email address will not be published. Required fields are marked *