artificial_intelligence_ai

Artificial Intelligence (AI)

Return to Deep learning, Machine Learning, Programming topics, Programming languages, Software engineering topics, Software architecture, Software architecture topics, Awesome lists, Azure AI, Azure Machine Learning services, Azure Batch AI

Artificial Intelligence (AI) refers to the simulation of human intelligence processes by machines, particularly computer systems. AI encompasses a wide range of techniques and methodologies aimed at enabling computers to perform tasks that typically require human intelligence, such as learning, reasoning, problem-solving, perception, understanding natural language, and decision-making. AI systems leverage various approaches, including machine learning, deep learning, natural language processing, computer vision, robotics, and expert systems, to emulate human-like intelligence and behavior. AI has diverse applications across industries, including healthcare, finance, transportation, education, entertainment, and more. Some common examples of AI applications include virtual assistants (such as Siri, Alexa, and Google Assistant), recommendation systems (like those used by streaming services and e-commerce platforms), autonomous vehicles, medical diagnosis systems, and fraud detection algorithms. As AI technologies continue to advance, they hold the potential to revolutionize how humans interact with technology and address complex problems, driving innovation and transforming various aspects of society and the economy.

Snippet from Wikipedia: Artificial intelligence

Artificial intelligence (AI) refers to the capability of computational systems to perform tasks typically associated with human intelligence, such as learning, reasoning, problem-solving, perception, and decision-making. It is a field of research in computer science that develops and studies methods and software that enable machines to perceive their environment and use learning and intelligence to take actions that maximize their chances of achieving defined goals. Such machines may be called AIs.

High-profile applications of AI include advanced web search engines (e.g., Google Search); recommendation systems (used by YouTube, Amazon, and Netflix); virtual assistants (e.g., Google Assistant, Siri, and Alexa); autonomous vehicles (e.g., Waymo); generative and creative tools (e.g., ChatGPT and AI art); and superhuman play and analysis in strategy games (e.g., chess and Go). However, many AI applications are not perceived as AI: "A lot of cutting edge AI has filtered into general applications, often without being called AI because once something becomes useful enough and common enough it's not labeled AI anymore."

Various subfields of AI research are centered around particular goals and the use of particular tools. The traditional goals of AI research include learning, reasoning, knowledge representation, planning, natural language processing, perception, and support for robotics. General intelligence—the ability to complete any task performed by a human on an at least equal level—is among the field's long-term goals. To reach these goals, AI researchers have adapted and integrated a wide range of techniques, including search and mathematical optimization, formal logic, artificial neural networks, and methods based on statistics, operations research, and economics. AI also draws upon psychology, linguistics, philosophy, neuroscience, and other fields.

Artificial intelligence was founded as an academic discipline in 1956, and the field went through multiple cycles of optimism throughout its history, followed by periods of disappointment and loss of funding, known as AI winters. Funding and interest vastly increased after 2012 when deep learning outperformed previous AI techniques. This growth accelerated further after 2017 with the transformer architecture, and by the early 2020s many billions of dollars were being invested in AI and the field experienced rapid ongoing progress in what has become known as the AI boom. The emergence of advanced generative AI in the midst of the AI boom and its ability to create and modify content exposed several unintended consequences and harms in the present and raised concerns about the risks of AI and its long-term effects in the future, prompting discussions about regulatory policies to ensure the safety and benefits of the technology.

External Sites

Terms related to: AI-ML-DL-NLP-GenAI-LLM-GPT-RAG-MLOps-Chatbots-ChatGPT-Gemini-Copilot-HuggingFace-GPU-Prompt Engineering-Data Science-DataOps-Data Engineering-Big Data-Analytics-Databases-SQL-NoSQL

AI, Artificial Intelligence (AI), Machine Learning (ML), Deep Learning (DL), Neural Network, Generative AI (GenAI), Natural Language Processing (NLP), Large Language Model (LLM), Transformer Models, GPT (Generative Pre-trained Transformer), ChatGPT, Chatbots, Prompt Engineering, HuggingFace, GPU (Graphics Processing Unit), RAG (Retrieval-Augmented Generation), MLOps (Machine Learning Operations), Data Science, DataOps (Data Operations), Data Engineering, Big Data, Analytics, Databases, SQL (Structured Query Language), NoSQL, Gemini (Google AI Model), Copilot (AI Pair Programmer), Foundation Models, LLM Fine-Tuning, LLM Inference, LLM Training, Parameter-Efficient Tuning, Instruction Tuning, Few-Shot Learning, Zero-Shot Learning, One-Shot Learning, Meta-Learning, Reinforcement Learning from Human Feedback (RLHF), Self-Supervised Learning, Contrastive Learning, Masked Language Modeling, Causal Language Modeling, Attention Mechanism, Self-Attention, Multi-Head Attention, Positional Embeddings, Word Embeddings, Tokenization, Byte Pair Encoding (BPE), SentencePiece Tokenization, Subword Tokenization, Prompt Templates, Prompt Context Window, Context Length, Scaling Laws, Parameter Scaling, Model Architecture, Model Distillation, Model Pruning, Model Quantization, Model Compression, Low-Rank Adaptation (LoRA), Sparse Models, Mixture of Experts, Neural Architecture Search (NAS), AutoML, Gradient Descent Optimization, Stochastic Gradient Descent (SGD), Adam Optimizer, AdamW Optimizer, RMSProp Optimizer, Adagrad Optimizer, Adadelta Optimizer, Nesterov Momentum, Learning Rate Schedules, Warmup Steps, Cosine Decay, Hyperparameter Tuning, Bayesian Optimization, Grid Search, Random Search, Population Based Training, Early Stopping, Regularization, Dropout, Weight Decay, Label Smoothing, Batch Normalization, Layer Normalization, Instance Normalization, Group Normalization, Residual Connections, Skip Connections, Encoder-Decoder Architecture, Encoder Stack, Decoder Stack, Cross-Attention, Feed-Forward Layers, Position-Wise Feed-Forward Network, Pre-LN vs Post-LN, Sequence-to-Sequence Models, Causal Decoder-Only Models, Masked Autoencoder, Domain Adaptation, Task-Specific Heads, Classification Head, Regression Head, Token Classification Head, Sequence Classification Head, Multiple-Choice Head, Span Prediction Head, Causal Head, Next Sentence Prediction, MLM (Masked Language Modeling), NSP (Next Sentence Prediction), C4 Dataset, WebText Dataset, Common Crawl Corpus, Wikipedia Corpus, BooksCorpus, Pile Dataset, LAION Dataset, Curated Corpora, Fine-Tuning Datasets, Instruction Data, Alignment Data, Human Feedback Data, Preference Ranking, Reward Modeling, RLHF Policy Optimization, Batch Inference, Online Inference, Vector Databases, FAISS Integration, Chroma Integration, Weaviate Integration, Pinecone Integration, Milvus Integration, Data Embeddings, Semantic Search, Embedding Models, Text-to-Vector Encoding, Vector Similarity Search, Approximate Nearest Neighbor (ANN), HNSW Index, IVF Index, ScaNN Index, Memory Footprint Optimization, HuggingFace Transformers, HuggingFace Hub, HuggingFace Datasets, HuggingFace Model Cards, HuggingFace Spaces, HuggingFace Inference Endpoints, HuggingFace Accelerate, HuggingFace PEFT (Parameter Efficient Fine-Tuning), HuggingFace Safetensors Format, HuggingFace Tokenizers, HuggingFace Pipeline, HuggingFace Trainer, HuggingFace Auto Classes (AutoModel, AutoTokenizer), HuggingFace Model Conversion, HuggingFace Community Models, HuggingFace Diffusers, Stable Diffusion, HuggingFace Model Hub Search, HuggingFace Secrets Management, OpenAI GPT models, OpenAI API, OpenAI Chat Completions, OpenAI Text Completions, OpenAI Embeddings API, OpenAI Rate Limits, OpenAI Fine-Tuning (GPT-3.5, GPT-4), OpenAI System Messages, OpenAI Assistant Messages, OpenAI User Messages, OpenAI Function Calls, OpenAI ChatML Format, OpenAI Temperature Parameter, OpenAI Top_p Parameter, OpenAI Frequency Penalty, OpenAI Presence Penalty, OpenAI Max Tokens Parameter, OpenAI Logit Bias, OpenAI Stop Sequences, Azure OpenAI Integration, Anthropic Claude Integration, Anthropic Claude Context Window, Anthropic Claude Constitutional AI, Cohere Integration LLM provider, Llama2 (Meta's LLM), Llama2 Chat Model, Vicuna Model (LLM)), Alpaca Model, StableLM, MPT (MosaicML Pretrained Transformer), Falcon LLM, Baichuan LLM, Code Llama, WizardCoder Model, WizardLM Model, Phoenix LLM, Samantha LLM, LoRA Adapters, PEFT for LLM, BitFit Parameters Tuning, QLoRA (Quantized LoRA), GLoRA, GGML Quantization, GPTQ Quantization, SmoothQuant, Int4 Quantization, Int8 Quantization, FP16 Mixed Precision, BF16 Precision, MLOps Tools, MLOps CI/CD, MLOps CD4ML, MLOps Feature Store, MLOps Model Registry, MLOps Model Serving, MLOps Model Monitoring, MLOps Model Drift Detection, MLOps Data Drift Detection, MLOps Model Explainability Integration, MLOps MLFlow Integration, MLOps Kubeflow Integration, MLOps MLRun, MLOps Seldon Core for serving, MLOps BentoML for serving, MLOps MLflow Tracking, MLOps MLflow Model Registry, MLOps DVC (Data Version Control), MLOps Delta Lake, RAG (Retrieval-Augmented Generation), RAG Document Store, RAG Vector Store Backend, RAG Memory Augmentation, RAG On-the-fly Retrieval, RAG Re-ranking Step, RAG HyDE Technique - It's known as hypothetical document embeddings - advanced but known in RAG, RAG chain-of-thought, chain-of-thought related to LLM reasoning, Chain-of-Thought Reasoning, Self-Consistency Decoding, Tree-of-thoughts, ReAct (Reason+Act) Prompting Strategy, Prompt Engineering Techniques, Prompt Templates (LLM), Prompt Variables Replacement, Prompt Few-Shot Examples, Prompt Zero-Shot Mode, Prompt Retrieval Injection, Prompt System Message, Prompt Assistant Message, Prompt Role Specification, Prompt Content Filtering, Prompt Moderation Tools, AI-Generated Code Completion, Copilot (GitHub) Integration, CoPilot CLI, Copilot Labs, Gemini (Google Model) Early access, LLM from Google, LaMDA (Language Model for Dialog Applications), PaLM (Pathways Language Model), PaLM2 (PaLM 2 Model), Flan PaLM Models, Google Vertex AI Integration, AWS Sagemaker Integration, Azure Machine Learning Integration, Databricks MLFlow Integration, HuggingFace Hub LFS for large models, LFS big files management, OPT (Open Pretrained Transformer) Meta Model, Bloom LLM, Ernie Bot (Baidu LLM), Zhipu-Chat - Another LLM from China, Salesforce CodeT5 - It's a code model, Finetune with LoRA on GPT-4, Anthropic Claude 2

Artificial Intelligence (AI): The Borg, SkyNet, Google Gemini, ChatGPT, AI Fundamentals, AI Inventor: Arthur Samuel of IBM 1959 coined term Machine Learning. Synonym Self-Teaching Computers from 1950s. Experimental AILearning Machine” called Cybertron in early 1960s by Raytheon Company; ChatGPT, Generative AI, NLP, GAN, AI winter, The Singularity, AI FUD, Quantum FUD (Fake Quantum Computers), AI Propaganda, Quantum Propaganda, Cloud AI (AWS AI, Azure AI, Google AI-GCP AI-Google Cloud AI, IBM AI, Apple AI), Deep Learning (DL), Machine learning (ML), AI History, AI Bibliography, Manning AI-ML-DL-NLP-GAN Series, AI Glossary, AI Topics, AI Courses, AI Libraries, AI frameworks, AI GitHub, AI Awesome List. (navbar_ai - See also navbar_dl, navbar_ml, navbar_nlp, navbar_chatbot, navbar_chatgpt, navbar_llm, navbar_openai, borg_usage_disclaimer, navbar_bigtech, navbar_cia)


Cloud Monk is Retired ( for now). Buddha with you. © 2025 and Beginningless Time - Present Moment - Three Times: The Buddhas or Fair Use. Disclaimers

SYI LU SENG E MU CHYWE YE. NAN. WEI LA YE. WEI LA YE. SA WA HE.


artificial_intelligence_ai.txt · Last modified: 2025/02/01 07:19 by 127.0.0.1

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki