Table of Contents
The Complete Obsolete Guide to Generative AI
by David Clinton
July 2024 ISBN 978-1633436985 240 pages
The last book on AI you’ll ever need. We swear!
AI technology moves so fast that this book is probably already out of date! But don’t worry — The Complete Obsolete Guide to Generative AI is still an essential read for anyone who wants to make gen AI into an AI tool rather than a AI toy. It shows you how to get the best out of AI no matter what changes come in the future. You’ll be able to use common automation and scripting tools to take AI to a new level, and access raw (and powerful) GPT models via ChatGPT API.
Inside The Complete Obsolete Guide to Generative AI you will find:
- Just enough background info on AI! What an AI model is how it works
- Ways to create text, code, and images for your organization's needs
- Training AI models on your local data stores or on the internet
- Business intelligence and analytics uses for AI
- Building your own custom AI models
- Looking ahead to the future of generative AI
Where to get started? How about creating exciting images, video, and even audio with AI. Need more? Learn to harness AI to speed up any everyday work task, including writing boilerplate code, creating specialized documents, and analyzing your own data. Push beyond simple ChatGPT prompts! Discover ways to double your productivity and take on projects you never thought were possible! AI—and this book—are here to show you how.
About the technology
Everything you learn about Generative AI tools like Chat-GPT, Copilot, and Claude becomes obsolete almost immediately. So how do you decide where to spend your time—and your company’s money? This entertaining and unbelievably practical book shows you what you can (and should!) do with AI now and how to roll with the changes as they happen.
About the book
The Complete Obsolete Guide to Generative AI is a lighthearted introduction to Generative AI written for technology professionals and motivated AI enthusiasts. In it, you’ll get a quick-paced survey of AI techniques for creating code, text, images, and presentations, working with data, and much more. As you explore the hands-on exercises, you’ll build an intuition for how Generative AI can transform your daily work and communication—and maybe even learn how to make peace with your new robot overlords.
What's inside
- The big picture of Generative AI tools and tech
- Creating useful text, code, and images
- Writing effective prompts
About the reader
Written for developers, admins, and other IT pros. Some examples use simple Python code.
About the author
David Clinton is an AWS Solutions Architect, a Linux server administrator and a world-renowned expert on obsolescence.
The technical editor on this book was Maris Sekar.
Terms related to: AI-ML-DL-NLP-GenAI-LLM-GPT-RAG-MLOps-Chatbots-ChatGPT-Gemini-Copilot-HuggingFace-GPU-Prompt Engineering-Data Science-DataOps-Data Engineering-Big Data-Analytics-Databases-SQL-NoSQL
AI, Artificial Intelligence (AI), Machine Learning (ML), Deep Learning (DL), Neural Network, Generative AI (GenAI), Natural Language Processing (NLP), Large Language Model (LLM), Transformer Models, GPT (Generative Pre-trained Transformer), ChatGPT, Chatbots, Prompt Engineering, HuggingFace, GPU (Graphics Processing Unit), RAG (Retrieval-Augmented Generation), MLOps (Machine Learning Operations), Data Science, DataOps (Data Operations), Data Engineering, Big Data, Analytics, Databases, SQL (Structured Query Language), NoSQL, Gemini (Google AI Model), Copilot (AI Pair Programmer), Foundation Models, LLM Fine-Tuning, LLM Inference, LLM Training, Parameter-Efficient Tuning, Instruction Tuning, Few-Shot Learning, Zero-Shot Learning, One-Shot Learning, Meta-Learning, Reinforcement Learning from Human Feedback (RLHF), Self-Supervised Learning, Contrastive Learning, Masked Language Modeling, Causal Language Modeling, Attention Mechanism, Self-Attention, Multi-Head Attention, Positional Embeddings, Word Embeddings, Tokenization, Byte Pair Encoding (BPE), SentencePiece Tokenization, Subword Tokenization, Prompt Templates, Prompt Context Window, Context Length, Scaling Laws, Parameter Scaling, Model Architecture, Model Distillation, Model Pruning, Model Quantization, Model Compression, Low-Rank Adaptation (LoRA), Sparse Models, Mixture of Experts, Neural Architecture Search (NAS), AutoML, Gradient Descent Optimization, Stochastic Gradient Descent (SGD), Adam Optimizer, AdamW Optimizer, RMSProp Optimizer, Adagrad Optimizer, Adadelta Optimizer, Nesterov Momentum, Learning Rate Schedules, Warmup Steps, Cosine Decay, Hyperparameter Tuning, Bayesian Optimization, Grid Search, Random Search, Population Based Training, Early Stopping, Regularization, Dropout, Weight Decay, Label Smoothing, Batch Normalization, Layer Normalization, Instance Normalization, Group Normalization, Residual Connections, Skip Connections, Encoder-Decoder Architecture, Encoder Stack, Decoder Stack, Cross-Attention, Feed-Forward Layers, Position-Wise Feed-Forward Network, Pre-LN vs Post-LN, Sequence-to-Sequence Models, Causal Decoder-Only Models, Masked Autoencoder, Domain Adaptation, Task-Specific Heads, Classification Head, Regression Head, Token Classification Head, Sequence Classification Head, Multiple-Choice Head, Span Prediction Head, Causal Head, Next Sentence Prediction, MLM (Masked Language Modeling), NSP (Next Sentence Prediction), C4 Dataset, WebText Dataset, Common Crawl Corpus, Wikipedia Corpus, BooksCorpus, Pile Dataset, LAION Dataset, Curated Corpora, Fine-Tuning Datasets, Instruction Data, Alignment Data, Human Feedback Data, Preference Ranking, Reward Modeling, RLHF Policy Optimization, Batch Inference, Online Inference, Vector Databases, FAISS Integration, Chroma Integration, Weaviate Integration, Pinecone Integration, Milvus Integration, Data Embeddings, Semantic Search, Embedding Models, Text-to-Vector Encoding, Vector Similarity Search, Approximate Nearest Neighbor (ANN), HNSW Index, IVF Index, ScaNN Index, Memory Footprint Optimization, HuggingFace Transformers, HuggingFace Hub, HuggingFace Datasets, HuggingFace Model Cards, HuggingFace Spaces, HuggingFace Inference Endpoints, HuggingFace Accelerate, HuggingFace PEFT (Parameter Efficient Fine-Tuning), HuggingFace Safetensors Format, HuggingFace Tokenizers, HuggingFace Pipeline, HuggingFace Trainer, HuggingFace Auto Classes (AutoModel, AutoTokenizer), HuggingFace Model Conversion, HuggingFace Community Models, HuggingFace Diffusers, Stable Diffusion, HuggingFace Model Hub Search, HuggingFace Secrets Management, OpenAI GPT models, OpenAI API, OpenAI Chat Completions, OpenAI Text Completions, OpenAI Embeddings API, OpenAI Rate Limits, OpenAI Fine-Tuning (GPT-3.5, GPT-4), OpenAI System Messages, OpenAI Assistant Messages, OpenAI User Messages, OpenAI Function Calls, OpenAI ChatML Format, OpenAI Temperature Parameter, OpenAI Top_p Parameter, OpenAI Frequency Penalty, OpenAI Presence Penalty, OpenAI Max Tokens Parameter, OpenAI Logit Bias, OpenAI Stop Sequences, Azure OpenAI Integration, Anthropic Claude Integration, Anthropic Claude Context Window, Anthropic Claude Constitutional AI, Cohere Integration LLM provider, Llama2 (Meta's LLM), Llama2 Chat Model, Vicuna Model (LLM)), Alpaca Model, StableLM, MPT (MosaicML Pretrained Transformer), Falcon LLM, Baichuan LLM, Code Llama, WizardCoder Model, WizardLM Model, Phoenix LLM, Samantha LLM, LoRA Adapters, PEFT for LLM, BitFit Parameters Tuning, QLoRA (Quantized LoRA), GLoRA, GGML Quantization, GPTQ Quantization, SmoothQuant, Int4 Quantization, Int8 Quantization, FP16 Mixed Precision, BF16 Precision, MLOps Tools, MLOps CI/CD, MLOps CD4ML, MLOps Feature Store, MLOps Model Registry, MLOps Model Serving, MLOps Model Monitoring, MLOps Model Drift Detection, MLOps Data Drift Detection, MLOps Model Explainability Integration, MLOps MLFlow Integration, MLOps Kubeflow Integration, MLOps MLRun, MLOps Seldon Core for serving, MLOps BentoML for serving, MLOps MLflow Tracking, MLOps MLflow Model Registry, MLOps DVC (Data Version Control), MLOps Delta Lake, RAG (Retrieval-Augmented Generation), RAG Document Store, RAG Vector Store Backend, RAG Memory Augmentation, RAG On-the-fly Retrieval, RAG Re-ranking Step, RAG HyDE Technique - It's known as hypothetical document embeddings - advanced but known in RAG, RAG chain-of-thought, chain-of-thought related to LLM reasoning, Chain-of-Thought Reasoning, Self-Consistency Decoding, Tree-of-thoughts, ReAct (Reason+Act) Prompting Strategy, Prompt Engineering Techniques, Prompt Templates (LLM), Prompt Variables Replacement, Prompt Few-Shot Examples, Prompt Zero-Shot Mode, Prompt Retrieval Injection, Prompt System Message, Prompt Assistant Message, Prompt Role Specification, Prompt Content Filtering, Prompt Moderation Tools, AI-Generated Code Completion, Copilot (GitHub) Integration, CoPilot CLI, Copilot Labs, Gemini (Google Model) Early access, LLM from Google, LaMDA (Language Model for Dialog Applications), PaLM (Pathways Language Model), PaLM2 (PaLM 2 Model), Flan PaLM Models, Google Vertex AI Integration, AWS Sagemaker Integration, Azure Machine Learning Integration, Databricks MLFlow Integration, HuggingFace Hub LFS for large models, LFS big files management, OPT (Open Pretrained Transformer) Meta Model, Bloom LLM, Ernie Bot (Baidu LLM), Zhipu-Chat - Another LLM from China, Salesforce CodeT5 - It's a code model, Finetune with LoRA on GPT-4, Anthropic Claude 2
Artificial Intelligence (AI): The Borg, SkyNet, Google Gemini, ChatGPT, AI Fundamentals, AI Inventor: Arthur Samuel of IBM 1959 coined term Machine Learning. Synonym Self-Teaching Computers from 1950s. Experimental AI “Learning Machine” called Cybertron in early 1960s by Raytheon Company; ChatGPT, Generative AI, NLP, GAN, AI winter, The Singularity, AI FUD, Quantum FUD (Fake Quantum Computers), AI Propaganda, Quantum Propaganda, Cloud AI (AWS AI, Azure AI, Google AI-GCP AI-Google Cloud AI, IBM AI, Apple AI), Deep Learning (DL), Machine learning (ML), AI History, AI Bibliography, Manning AI-ML-DL-NLP-GAN Series, AI Glossary, AI Topics, AI Courses, AI Libraries, AI frameworks, AI GitHub, AI Awesome List. (navbar_ai - See also navbar_dl, navbar_ml, navbar_nlp, navbar_chatbot, navbar_chatgpt, navbar_llm, navbar_openai, borg_usage_disclaimer, navbar_bigtech, navbar_cia)
ChatGPT: The Borg, ChatGPT Alternatives (Google Gemini, Microsoft Copilot, Cursor AI, JetBrains AI), Sora.com, ChatGPT Pro, ChatGPT Plus, GPT - Chat - Chatbot, OpenAI, ChatGPT Prompts, ChatGPT o1-Pro, ChatGPT o1, ChatGPT‑4o, ChatGPT 4.0, ChatGPT 3.5, ChatGPT Dynamic, Prompt Engineering, Chatbots, AI Assistants, Chat Generative Pre-trained Transformer developed by OpenAI and launched on November 30, 2022; Large Language Model (LLM), Language model, Prompt Engineering, Generative Artificial Intelligence (GenAI), Generative pre-trained transformer, GPT-4, GPT-3, GPT-2, Brave Leo, Transformer (machine learning model), Fine-tuning (deep learning), Supervised learning (SL), Microsoft Bing, Bing Chat, LLaMA, Artificial neural networks (ANNs, also shortened to neural networks (NNs) or neural nets), Generative model, AI accelerator, Fine-tuning (machine learning) –> Fine-tuning (deep learning), Transfer learning (TL), Multimodal learning, GitHub ChatGPT, Awesome ChatGPT. (navbar_chatgpt - see also navbar_ml, navbar_dl, navbar_nlp, navbar_ai, navbar_llm, navbar_openai, borg_usage_disclaimer, navbar_cia)
Manning Publications: Manning Books Purchased by Cloud Monk, Manning Books Series, Manning Bibliography, In a Month of Lunches, In Action, Manning API Series, Manning "Functional Programming in" Series, Manning Concurrency Async Multithreaded Parallel Programming Series, Manning Grokking Series, Manning Java-JVM Languages Series (Manning Java Series, Manning Kotlin Series), Manning JavaScript Series, Manning TypeScript Series, Manning Microservices Series, Manning Python Series, Manning Security Series, Manning Spring Series, Manning SQL Series, Manning Database Series, Manning Data Science Series, Manning Mistakes and How to Avoid Them Series, Manning Books that were Cancelled, MEAP, Cloud Monk's Book Purchases, Cloud Monk Library. (navbar_manning)
Cloud Monk is Retired ( for now). Buddha with you. © 2025 and Beginningless Time - Present Moment - Three Times: The Buddhas or Fair Use. Disclaimers
SYI LU SENG E MU CHYWE YE. NAN. WEI LA YE. WEI LA YE. SA WA HE.