GenAI
Generative artificial intelligence (generative AI, GenAI, or GAI) is a subset of artificial intelligence that uses generative models to produce text, images, videos, or other forms of data. These models often generate output in response to specific prompts. Generative AI systems learn the underlying patterns and structures of their training data, enabling them to create new data.
See:
Resources
- Generative artificial intelligence - Wikipedia
- Driving Enterprise Transformation With Generative AI | Dataiku
- The state of GenAI in the enterprise (Deloitte)
- ¿Qué es la IA generativa? - Explicación de la inteligencia artificial generativa - AWS (amazon.com)
- La IA Generativa Abre una Nueva Era de Eficiencia en Todas las Industrias | Blog de NVIDIA
- IA Generativa - Capgemini Spain
- El estado de la IA en 2023: El año clave de la IA generativa | McKinsey
- La IA generativa en los negocios | Accenture
- Modelos IA: riesgos y oportunidades - KPMG España
- ¿Qué es la inteligencia artificial generativa? Ejemplos y riesgos (redhat.com)
- ¿Qué es la IA generativa? ¿Cómo funciona? | Oracle España
- The ultimate AI glossary to help you navigate our changing world | Mashable
- Microsoft Source: 10 AI terms everyone should know - 10 AI terms
- Artificial Intelligence (AI) Terms: A to Z Glossary | Coursera
- Glossary of artificial intelligence - Wikipedia
Aplicationts per sector
- La IA generativa por sectores - KPMG Tendencias
- Generative AI for Energy Sector and its Benefits (xenonstack.com)
- Generative AI in energy, natural resources, and chemicals (kpmg.com)
- La IA en el sector energético: Principales casos de uso - GAMCO, SL
Applications to science
See AI for scientific discovery
Generative modelling
- Generative models
- Deep Generative Models
- Taxonomy of Generative Models
- Jakub Tomczak blog:
Foundation models
- A foundation model is a large artificial intelligence model trained on a vast quantity of unlabeled data at scale (usually by AI/Self-supervised learning) resulting in a model that can be adapted to a wide range of downstream tasks
- https://en.wikipedia.org/wiki/Foundation_models
- https://en.wikipedia.org/wiki/Large_language_model
- Center for Research on Foundation Models (CRFM)
- Foundation Models and the Future of Multi-Modal AI
- Foundation models: 2022’s AI paradigm shift
- Foundation Models: paradigm shift for AI or mere rebranding?
- ChatGPT, LLMs, and Foundation models — a closer look into the hype and implications for startups
Benchmarks
Courses
- #COURSE Driving Innovation with Generative AI | MIT xPRO
- #COURSE Deep Generative Modeling: VAEs and GANs (MIT 6.S191)
- #COURSE Deep Generative Models (Stanford CS236 - Fall 2021)
- #COURSE Deep Generative Models lecture (Carnegie Mellon University)
- #COURSE Generative Deep Learning with TensorFlow (Deeplearning.AI coursera)
Code
Books
Subtopics
Variational Autoencoders
See AI/Deep Learning/Autoencoders#VAEs
GANs
Normalizing flows
See AI/Deep Learning/Normalizing flows
Diffusion models
See AI/Deep Learning/Diffusion models
GFlow nets
See GFlowNets
Generative models for computer vision
See:
- VLMs
- Diffusion models
- Multimodal learning
- AI/Computer Vision/Image-to-image translation#GAN-based
- AI/Deep Learning/GANs#GANs for representation learning and image synthesis
- AI/Deep Learning/Transformers#For Computer Vision
- Deep Computer Vision
Generative AI for time series
See AI/Generative AI/GenAI for time series
Generative AI for tabular data
Generative models for audio
See GenAI for audio
Generative AI for text and NLP
See LLMs
References
- #PAPER On the Opportunities and Risks of Foundation Models (Bommasani 2021)
- A foundation model is any model that is trained on broad data at scale and can be adapted (e.g., fine-tuned) to a wide range of downstream tasks; current examples include BERT, GPT-3, and CLIP
- Foundation models are based on deep neural networks and self-supervised learning
- On a technical level, foundation models are enabled by transfer learning and scale
- The idea of transfer learning is to take the “knowledge” learned from one task (e.g., object recognition in images) and apply it to another task (e.g., activity recognition in videos).
- Within deep learning, pretraining is the dominant approach to transfer learning: a model is trained on a surrogate task (often just as a means to an end) and then adapted to the downstream task of interest via fine-tuning
- Transfer learning is what makes foundation models possible, but scale is what makes them powerful. Scale required three ingredients: improvements in computer hardware, the development of the Transformer model architecture that leverages the parallelism of the hardware to train much more expressive models than before and the availability of much more training data
- #PAPER #REVIEW ChatGPT is not all you need. A State of the Art Review of large Generative AI models (Gozalo-Brizuela 2023)