Reference Guide
100 AI and LLM Questions Answered
Fact-based answers to the most important questions about artificial intelligence, large language models, prompting, safety, and business use cases. Every answer reflects consensus across major AI providers and researchers.
100 questions across 10 categories · External sources linked throughout
1AI Fundamentals
Q1.What is artificial intelligence (AI)?
Artificial intelligence (AI) refers to computer systems that perform tasks typically requiring human intelligence, such as understanding language, recognizing patterns, and making decisions. Modern AI is primarily based on machine learning, where systems learn from large amounts of data rather than following explicitly programmed rules.
IBM: What is AI?Q2.What is a large language model (LLM)?
A large language model (LLM) is a type of AI model trained on vast amounts of text data to understand and generate human language. LLMs learn statistical patterns across billions of text examples, enabling them to answer questions, write content, summarize documents, translate languages, and assist with coding. Examples include GPT-4, Claude, Gemini, and Llama.
Survey of LLMs (arXiv)Q3.What is machine learning?
Machine learning is a subset of AI where systems learn to perform tasks by finding patterns in data, rather than being explicitly programmed with rules. A model is trained on examples, adjusts its internal parameters, and then applies what it has learned to new, unseen inputs. Deep learning, a subset of machine learning using multi-layered neural networks, powers most modern AI applications.
Q4.What is the difference between AI and machine learning?
AI is the broader field concerned with building systems that can perform intelligent tasks. Machine learning is one approach to achieving AI, where systems learn from data rather than following hand-written rules. Today, most advanced AI systems are built on machine learning, but the terms are not interchangeable.
Q5.What is generative AI?
Generative AI refers to AI systems that create new content, including text, images, audio, video, and code. Unlike traditional AI that classifies or predicts from existing data, generative AI produces novel outputs. Large language models, image generators like DALL-E and Midjourney, and audio models like ElevenLabs are all examples of generative AI.
Q6.What is a neural network?
A neural network is a computational model loosely inspired by the structure of the human brain, consisting of layers of interconnected nodes that process and transform data. Neural networks learn by adjusting the strength of connections between nodes based on training examples. Deep neural networks with many layers form the foundation of modern AI systems including LLMs.
Q7.What is a transformer model?
A transformer is a neural network architecture introduced in the 2017 paper "Attention Is All You Need" that revolutionized natural language processing. Transformers use a mechanism called self-attention to weigh the relevance of different parts of a sequence relative to each other. Virtually all modern LLMs, including GPT-4, Claude, and Gemini, are built on transformer architectures.
"Attention Is All You Need" (arXiv)Q8.What is natural language processing (NLP)?
Natural language processing (NLP) is the branch of AI focused on enabling computers to understand, interpret, and generate human language. NLP includes tasks such as text classification, translation, summarization, sentiment analysis, and question answering. Modern NLP is dominated by LLMs that handle many of these tasks within a single model.
Q9.What is training data?
Training data is the dataset used to teach a machine learning model. For LLMs, training data typically consists of hundreds of billions of tokens of text sourced from websites, books, code repositories, and other text sources. The quality, diversity, and scale of training data significantly influence a model's capabilities and any biases it may exhibit.
Q10.What is a foundation model?
A foundation model is a large AI model trained on broad data at scale that can be adapted for a wide range of downstream tasks. Foundation models like GPT-4 and Claude are trained once at great cost and then fine-tuned or prompted for specific applications. The term was introduced by Stanford's Center for Research on Foundation Models.
Stanford CRFM2How LLMs Work
Q11.How does an LLM generate text?
LLMs generate text by predicting the most likely next token (word or word fragment) given the preceding context, one token at a time. This process is called autoregressive generation. The model samples from a probability distribution over all possible tokens to produce coherent and contextually appropriate responses.
Q12.What are tokens in LLMs?
Tokens are the basic units that LLMs use to process text. A token is typically a word, part of a word, or a punctuation character, depending on the model's tokenizer. For English text, one token is roughly four characters or three-quarters of a word on average. LLM pricing and context window sizes are both measured in tokens.
Q13.What is a context window?
A context window is the maximum amount of text an LLM can process in a single interaction, measured in tokens. Everything the model "sees," including the system prompt, conversation history, and the current input, must fit within the context window. Larger context windows allow the model to consider more information at once.
Q14.What is prompt engineering?
Prompt engineering is the practice of designing and refining inputs (prompts) to elicit desired outputs from an AI model. Effective prompts provide clear instructions, relevant context, examples, and constraints that guide the model toward accurate and useful responses. Prompt engineering is a key skill for getting reliable results from LLMs.
Q15.What is fine-tuning?
Fine-tuning is the process of further training a pre-trained foundation model on a smaller, task-specific dataset to improve its performance on a particular domain or use case. Fine-tuning adjusts the model's weights to specialize its behavior without training from scratch. It is commonly used to make models follow specific formats, adopt a particular tone, or excel at domain-specific tasks.
Q16.What is Retrieval-Augmented Generation (RAG)?
Retrieval-Augmented Generation (RAG) is a technique that combines an LLM with a retrieval system to provide the model with relevant external information at inference time. Instead of relying solely on knowledge encoded in its parameters, a RAG system fetches relevant documents and includes them in the prompt context. This reduces hallucinations and allows the model to reference up-to-date or proprietary information.
Original RAG paper (arXiv)Q17.What is temperature in LLM outputs?
Temperature is a parameter that controls the randomness of an LLM's output. A temperature of 0 makes the model deterministic, always choosing the most probable next token. Higher temperatures introduce more variety and creativity, but also increase the risk of incoherence. Most LLM APIs allow users to set temperature between 0 and 1 (or sometimes higher).
Q18.What are embeddings?
Embeddings are numerical vector representations of text, images, or other data that capture semantic meaning in a high-dimensional space. Items with similar meanings are positioned close together in embedding space. Embeddings are used for semantic search, recommendation, clustering, and as the foundation of RAG systems and vector databases.
Q19.What is RLHF (Reinforcement Learning from Human Feedback)?
Reinforcement Learning from Human Feedback (RLHF) is a training technique used to align LLMs with human preferences. Human raters evaluate model outputs, and those ratings train a reward model that guides further fine-tuning of the LLM. RLHF is a key technique behind the helpful and safe behavior of models like ChatGPT, Claude, and Gemini.
RLHF paper (arXiv)Q20.What is the difference between model parameters and tokens?
Parameters are the internal numerical weights of a neural network learned during training that encode the model's knowledge. A model described as having 70 billion parameters contains 70 billion individual numerical values. Tokens are units of text processed during use. Parameters are a property of the trained model; tokens are units of input and output during inference.
3Capabilities
Q21.What can LLMs do well?
LLMs excel at tasks involving language: writing, summarizing, translating, explaining, brainstorming, classifying, extracting structured data from unstructured text, answering questions, and generating code. They perform particularly well on tasks where the format of a correct answer is learnable from patterns in training data, and they improve substantially when given clear, detailed prompts.
Q22.What is AI code generation?
AI code generation is the ability of LLMs to write functional code in programming languages like Python, JavaScript, SQL, and many others. Models can generate code from natural language descriptions, complete partially written functions, explain existing code, fix bugs, and write tests. Tools like GitHub Copilot, Cursor, and Claude Code are built on this capability.
Q23.Can LLMs reason?
LLMs can perform many tasks that require reasoning, such as solving math problems, answering multi-step logic questions, and planning sequences of actions. However, LLM reasoning emerges from pattern matching rather than formal logic, and models can make systematic errors on problems requiring precise step-by-step deduction. Techniques like chain-of-thought prompting significantly improve reasoning performance.
Q24.Do LLMs have access to the internet by default?
Standard LLMs do not have real-time internet access. Their knowledge comes from training data with a fixed cutoff date. Some products built on LLMs (such as Perplexity, ChatGPT with browsing, and Gemini with Search) integrate web search as an external tool, but this is an add-on feature, not a property of the base model.
Q25.What is multimodal AI?
Multimodal AI refers to models that can process and generate multiple types of data, such as text, images, audio, and video. A multimodal LLM like GPT-4o or Gemini can accept image inputs alongside text, analyze visual content, and respond with text. Multimodal capabilities allow AI to understand richer real-world inputs beyond text alone.
Q26.What is AI text summarization?
AI text summarization uses LLMs to condense long documents into shorter versions capturing the key points. Summarization can be extractive (selecting key sentences from the original) or abstractive (generating new text that captures the meaning). Modern LLMs perform abstractive summarization well, handling documents up to the length of their context window.
Q27.Can LLMs translate languages?
Yes. Modern LLMs are highly capable at translation between major world languages, often matching or exceeding specialized translation systems for common language pairs. They handle tone, nuance, and context better than earlier statistical translation models. Accuracy decreases for lower-resource languages with less training data.
Q28.What is sentiment analysis with AI?
Sentiment analysis is the use of AI to classify the emotional tone of text, typically as positive, negative, or neutral. LLMs can perform nuanced sentiment analysis beyond simple polarity, identifying specific emotions, assessing intensity, and handling sarcasm more reliably than earlier NLP methods. It is widely used for analyzing customer reviews, social media, and support tickets.
Q29.Can LLMs generate images?
Standard text-based LLMs do not generate images. Image generation requires separate generative models such as DALL-E 3 (OpenAI), Imagen (Google), or Stable Diffusion (Stability AI), which use different architectures such as diffusion models. Some products integrate both LLMs and image models to create multimodal experiences.
Q30.How accurate are LLMs at answering factual questions?
LLM accuracy on factual questions varies significantly by domain and specificity. Models are generally reliable on well-documented topics with extensive training data, but can hallucinate on niche, recent, or highly specific questions. Grounding responses in retrieved source documents (RAG) substantially improves factual accuracy.
4Limitations and Risks
Q31.What are AI hallucinations?
AI hallucinations occur when an LLM generates text that sounds confident and plausible but is factually incorrect or entirely fabricated. Hallucinations happen because LLMs are trained to produce statistically likely sequences of text, not to verify factual accuracy against ground truth. They are a fundamental limitation of current LLM architectures and a key reason why AI outputs should be verified for high-stakes uses.
Q32.Why do LLMs make factual errors?
LLMs make factual errors because they are trained to predict plausible text, not to retrieve verified facts. They compress knowledge into parameters imperfectly, and some information is learned unreliably. Training data also contains errors and contradictions that models may reproduce. Techniques like RAG and grounding in external sources reduce but do not eliminate factual errors.
Q33.What is a knowledge cutoff date?
A knowledge cutoff date is the point in time beyond which an LLM has no knowledge of world events, because no training data was collected after that date. A model with a cutoff of early 2024 will not know about events that occurred afterward unless that information is provided in the prompt. Providers publish cutoff dates for their models.
Q34.What is AI bias?
AI bias refers to systematic and unfair skews in AI outputs related to characteristics such as race, gender, age, or nationality. Bias in AI often originates from imbalances or stereotypes present in training data, and can manifest as different quality of service, harmful stereotypes, or unequal outcomes for different groups. Addressing AI bias is an active area of research and policy.
Q35.Should AI replace human judgment in important decisions?
Current AI systems should not replace human judgment in high-stakes decisions. LLMs can assist, inform, and accelerate human decision-making, but they lack lived experience, accountability, and the ability to verify their own reasoning reliably. Domains such as medicine, law, finance, and safety-critical systems require human oversight of AI-generated recommendations.
Q36.What is prompt injection?
Prompt injection is an attack where malicious instructions are embedded in content that an LLM processes, such as a document, email, or webpage, causing the model to follow those instructions instead of the user's intended task. It is a significant security concern for AI systems that process external content, analogous to SQL injection in traditional software. Active research exists on defenses, but no complete solution is currently available.
Q37.What are the privacy risks of using AI?
When users submit text to commercial AI services, that text may be stored, reviewed, or used for training depending on the provider's data practices. Sensitive information including personal details, passwords, or proprietary business data should not be entered into AI systems unless the service has strong, verified privacy guarantees. Enterprise AI agreements and on-premises deployments can mitigate these risks.
Q38.Can LLMs be confidently wrong?
Yes. LLMs frequently express incorrect information with the same confident, fluent tone they use for accurate information. The model has no reliable internal mechanism for knowing what it does not know. This is why critical-use AI outputs must be verified against authoritative sources, especially in medical, legal, and financial contexts.
Q39.What are the environmental costs of AI?
Training large AI models requires significant computational resources and energy. The training of GPT-3, for example, was estimated to emit hundreds of tonnes of CO2. Inference (running the model) at scale across millions of users also has a meaningful energy footprint. Major AI providers are investing in renewable energy and more efficient model architectures to address these costs.
Q40.What is AI toxicity?
AI toxicity refers to the generation of harmful, offensive, or abusive content by AI systems. LLMs trained on internet data can reproduce or generate harmful material without safeguards. AI providers use techniques including RLHF, content filtering, and safety fine-tuning to reduce toxicity, but no system eliminates it entirely.
5Prompting Best Practices
Q41.What makes a good AI prompt?
A good prompt is clear, specific, and provides the context necessary for the model to understand the task. Effective prompts include a clear description of the desired output, any relevant constraints or format requirements, background context, and examples where helpful. The more precisely you describe what you want, the more reliably you will get it.
Q42.What is chain-of-thought prompting?
Chain-of-thought (CoT) prompting is a technique where the user asks the model to reason through a problem step by step before giving a final answer. This approach improves performance on complex reasoning tasks like math, logic, and multi-step problems. Adding "think step by step" or "reason through this carefully" to a prompt can invoke chain-of-thought reasoning.
Chain-of-Thought Prompting paper (arXiv)Q43.What is few-shot prompting?
Few-shot prompting involves providing the model with a small number of examples (typically 2 to 5) of the desired input-output format before asking it to complete a new task. The examples help the model understand the pattern or format required without explicit instruction. Few-shot prompting is particularly effective for structured output tasks like classification, extraction, and formatting.
Q44.What is zero-shot prompting?
Zero-shot prompting means asking an LLM to perform a task without providing any examples. The model relies entirely on its training to understand and complete the request. Modern LLMs like GPT-4 and Claude are capable zero-shot learners for many tasks, requiring only clear natural-language instructions.
Q45.What is role prompting?
Role prompting assigns a persona or role to the LLM at the start of a conversation, for example "You are an expert tax accountant." This primes the model to draw on relevant knowledge and adopt an appropriate tone for the assigned role. The model's actual capabilities do not change, but role prompting shapes framing, vocabulary, and emphasis.
Q46.How do you reduce hallucinations through prompting?
To reduce hallucinations, provide source documents in the prompt context (RAG), ask the model to cite specific sources, instruct it to say "I don't know" when uncertain, and break complex questions into smaller verifiable steps. Setting temperature to 0 for factual tasks and requesting that the model explain its reasoning also help reduce confident errors.
Q47.What is prompt chaining?
Prompt chaining is the practice of breaking a complex task into a sequence of simpler prompts where the output of one becomes the input to the next. This approach is more reliable than asking a single prompt to accomplish an overly complex task. Prompt chaining is the foundation of many AI agent and workflow automation systems.
Q48.What is a system prompt?
A system prompt is an initial instruction provided to an LLM before the user's input that sets the model's behavior, persona, and constraints for the entire conversation. API users can define system prompts to customize how the model responds to all subsequent messages. Many enterprise AI applications use system prompts to scope the model's behavior to a specific use case.
Q49.How do you get consistent outputs from an LLM?
Consistency is improved by setting temperature to 0, using precise and detailed prompts, providing examples of desired output format, and requesting structured outputs like JSON where possible. For production systems, maintaining version-controlled prompts, testing across representative inputs, and monitoring output quality over time helps sustain consistency as models evolve.
Q50.What is structured output in LLMs?
Structured output refers to the ability to instruct an LLM to respond in a specific machine-readable format, such as JSON or XML, rather than free-form prose. Most major LLM APIs support structured output or JSON mode, which constrains the model to produce valid structured responses. Structured outputs are essential for building reliable AI pipelines and integrations.
6Business Use Cases
Q51.How can AI help with customer service?
AI can automate responses to common customer questions, route support tickets, summarize customer history for agents, and draft responses for human review. LLMs enable 24/7 availability and can handle high volumes of routine inquiries at low cost. The best implementations combine AI automation for routine tasks with human handoff for complex or sensitive issues.
Q52.How can businesses use AI for content creation?
Businesses use AI to draft blog posts, social media content, email campaigns, product descriptions, and internal documents. AI dramatically reduces the time from brief to first draft and allows rapid iteration. Human review and editing remain important for accuracy, brand voice, and quality control.
Q53.What is AI workflow automation?
AI workflow automation uses AI models to execute multi-step business processes that previously required human intervention. Examples include automated data extraction from documents, email triage and response drafting, lead qualification, and report generation. When combined with integration platforms, AI can act as an intelligent layer across a company's existing software stack.
Q54.How can AI improve sales processes?
AI can help sales teams by researching prospects, drafting personalized outreach, summarizing call transcripts, scoring leads, updating CRM records, and identifying upsell opportunities from customer data. AI-assisted sales tools let reps focus more time on conversations while reducing administrative overhead.
Q55.Can AI help with financial analysis?
AI can assist with financial analysis by extracting data from documents, summarizing financial reports, building models from natural language instructions, and identifying anomalies in transaction data. AI outputs in financial contexts require careful human review, as errors can have significant consequences. AI is most valuable as an accelerant for analyst workflows, not a replacement for professional judgment.
Q56.How can small businesses use AI?
Small businesses can use AI to write marketing copy, respond to customer inquiries, manage social media, draft contracts and standard documents, and automate repetitive back-office tasks. Consumer tools like ChatGPT and Claude require no technical expertise and offer significant time savings at low cost. The highest-value applications are usually those that eliminate repetitive, predictable work.
Q57.What is AI-powered search?
AI-powered search uses LLMs to understand the intent behind a query and return synthesized answers rather than just a list of links. Systems like Perplexity AI, Google AI Overviews, and Bing Copilot combine web search with LLM synthesis to provide direct answers with citations. AI search is changing how users find information and has significant implications for content and SEO strategy.
Q58.How can AI help with HR and recruiting?
AI can help HR teams by screening and ranking resumes, drafting job descriptions, generating interview questions, summarizing candidate profiles, and answering employee policy questions. AI tools must be used carefully in hiring to avoid amplifying bias; human review of AI-assisted screening decisions is essential and in many jurisdictions legally required.
Q59.What is AI-assisted decision making?
AI-assisted decision making uses AI to analyze data, surface patterns, generate options, and provide recommendations that inform human decisions. The human remains responsible for the final decision. Examples include credit risk scoring, medical diagnosis support, supply chain optimization, and fraud detection. The core principle is human oversight of consequential decisions.
Q60.How is AI used in marketing?
AI is used in marketing for content generation, audience segmentation, personalization, A/B test analysis, ad copy optimization, and customer journey mapping. AI tools allow marketing teams to produce more personalized content at scale and make faster, data-driven decisions. The highest-value applications connect AI to proprietary customer and performance data.
7Safety and Ethics
Q61.What is AI safety?
AI safety is the research field focused on ensuring that AI systems behave as intended and do not cause unintended harm. It encompasses near-term concerns (reducing harmful outputs, preventing misuse) and longer-term concerns (ensuring advanced AI remains aligned with human values as capabilities increase). Major AI labs including Anthropic, OpenAI, Google DeepMind, and Meta have dedicated AI safety research teams.
Q62.What is responsible AI?
Responsible AI refers to the principles and practices for developing and deploying AI in ways that are fair, transparent, accountable, safe, and beneficial to society. Responsible AI frameworks typically address bias mitigation, explainability, human oversight, privacy, and security. Organizations including Microsoft, Google, IBM, and the OECD have published responsible AI principles.
OECD AI PrinciplesQ63.What are AI guardrails?
AI guardrails are technical and policy mechanisms that limit what an AI system will say or do, preventing harmful, unethical, or inappropriate outputs. Guardrails include safety fine-tuning (RLHF), content filtering, output classifiers, and system prompt restrictions. All major AI providers implement guardrails, though their scope and strength vary by provider and use case.
Q64.What is the EU AI Act?
The EU AI Act is a comprehensive regulatory framework for artificial intelligence in the European Union that became law in 2024. It classifies AI systems by risk level (unacceptable, high, limited, and minimal risk) and imposes requirements proportional to risk, including transparency, human oversight, and documentation requirements. It is the first comprehensive AI law enacted by a major jurisdiction.
EU AI Act overviewQ65.What is AI alignment?
AI alignment is the challenge of ensuring AI systems pursue goals and exhibit behavior consistent with human values and intentions. A misaligned AI might optimize for a measurable proxy goal in ways that produce harmful side effects. Alignment is considered a central research challenge by major AI safety organizations, particularly as models become more capable and autonomous.
Q66.What is a deepfake?
A deepfake is synthetic media, typically video or audio, generated by AI to realistically depict a person saying or doing something they did not actually say or do. Deepfakes are created using generative AI models trained on real media of the target person. They raise significant concerns about disinformation, fraud, and non-consensual use, and have prompted legislation in many jurisdictions.
Q67.Who owns content generated by AI?
The legal status of AI-generated content ownership varies by jurisdiction and is still evolving. In the United States, the Copyright Office has ruled that purely AI-generated works without meaningful human authorship are not eligible for copyright protection. Content involving substantial human creative input in the prompting or editing process may qualify for protection. Users should review the terms of service of their AI provider.
Q68.What are the main ethical concerns with AI?
Key ethical concerns include: bias and discrimination in AI outputs, erosion of privacy through data collection, workforce displacement, spread of misinformation via AI-generated content, misuse for surveillance and manipulation, lack of transparency in AI decision-making, and environmental impact from training large models. These concerns are actively addressed by researchers, policymakers, and AI developers.
Q69.What is AI transparency?
AI transparency refers to the degree to which the behavior, capabilities, limitations, and decision-making of an AI system can be understood and audited by users and affected parties. Transparency measures include model cards (documentation of capabilities and limitations), explainability tools, and disclosure requirements. Transparency is a cornerstone of responsible AI deployment.
Q70.What is Constitutional AI?
Constitutional AI is a training approach developed by Anthropic in which an AI model is trained to follow a set of explicitly stated principles (a "constitution") rather than relying solely on human feedback for every behavior. The model critiques and revises its own outputs against the constitution during training. This approach is designed to make AI alignment more transparent and scalable.
Constitutional AI paper (Anthropic)8Leading AI Models and Providers
Q71.What is GPT-4?
GPT-4 is a large multimodal language model developed by OpenAI, released in March 2023. It can process both text and images and is capable of advanced reasoning, complex instruction following, and high-quality text generation across many domains. GPT-4 and its successors power ChatGPT Plus and the OpenAI API.
OpenAI GPT-4 researchQ72.What is Claude?
Claude is a large language model family developed by Anthropic, designed with a strong emphasis on safety, helpfulness, and honesty. Claude models are trained using Constitutional AI, a technique developed by Anthropic to align AI behavior with a set of stated principles. Claude is available via the Anthropic API and the Claude.ai web and mobile interfaces.
Anthropic ClaudeQ73.What is Gemini?
Gemini is Google's family of multimodal AI models designed to process text, images, audio, video, and code. Gemini models power Google's AI products including Search AI Overviews, Google Workspace features, and the Gemini chatbot. The family ranges from Gemini Nano (for on-device use) to Gemini Ultra (for complex reasoning tasks).
Google GeminiQ74.What is Llama?
Llama is a family of open-weight large language models developed by Meta AI. Unlike proprietary models, Llama model weights are publicly released, allowing researchers and developers to download, run, and fine-tune the models. Llama models have become the leading open-weight foundation models, widely used for research and commercial applications.
Meta LlamaQ75.What is Mistral AI?
Mistral AI is a French AI company founded in 2023 that has released several highly efficient open-weight language models. Mistral's models are notable for delivering strong performance relative to their size, making them popular for developers who want capable, cost-effective models that can be run locally or via API.
Mistral AIQ76.What is the difference between proprietary and open-weight AI models?
Proprietary AI models (such as GPT-4, Claude, and Gemini) are owned by companies that do not release model weights. Access is only available via API or consumer products, and the provider controls safety measures and pricing. Open-weight models (such as Llama and Mistral) release the model weights publicly, enabling local deployment, customization, and research without the managed safety infrastructure of proprietary APIs.
Q77.What is OpenAI?
OpenAI is an AI research company founded in 2015, creator of the GPT series of language models, the DALL-E image generation models, and the Whisper speech recognition system. OpenAI operates ChatGPT and the OpenAI API and has a capped-profit structure with Microsoft as its primary investor.
OpenAIQ78.What is Anthropic?
Anthropic is an AI safety company founded in 2021 by former OpenAI researchers including Dario and Daniela Amodei. Anthropic develops the Claude family of models, with a research focus on AI safety and alignment. Anthropic is backed by major investors including Amazon and Google.
AnthropicQ79.What is Google DeepMind?
Google DeepMind is Google's primary AI research organization, formed in 2023 by merging Google Brain and DeepMind. It is responsible for developing the Gemini model family, as well as landmark AI research including AlphaFold (protein structure prediction) and AlphaGo (the first AI to defeat a world champion Go player).
Google DeepMindQ80.What is an AI API?
An AI API (Application Programming Interface) is a service that allows developers to access AI model capabilities programmatically by sending requests with inputs and receiving AI-generated outputs. APIs from OpenAI, Anthropic, and Google allow developers to build AI-powered applications without managing the underlying model infrastructure. Pricing is typically based on tokens processed.
9Implementation and Costs
Q81.How much does it cost to use ChatGPT?
ChatGPT is available in a free tier providing access to GPT-4o mini and limited access to GPT-4o. ChatGPT Plus costs $20 per month and provides priority access to the latest GPT-4 models. For API access, pricing is charged per million tokens and varies by model. Pricing changes frequently as models improve and compute costs decrease.
OpenAI pricingQ82.How is LLM API usage priced?
LLM API pricing is typically charged separately for input tokens (what you send to the model) and output tokens (what the model generates). Costs vary significantly by model capability and provider. Smaller, faster models cost a fraction of a cent per thousand tokens, while the most capable models cost several cents per thousand tokens. Most providers offer a free tier for experimentation.
Q83.What is AI inference?
AI inference is the process of running a trained AI model to generate predictions or outputs in response to new inputs. Inference is distinct from training (which creates the model) and is what happens every time you send a message to an AI chatbot or call an AI API. Inference cost, latency, and energy consumption are key factors in the economics of AI products.
Q84.What is a vector database?
A vector database is a specialized database designed to store and efficiently search high-dimensional vector embeddings. Vector databases are a core component of RAG systems, enabling fast semantic search across large document collections. Popular vector databases include Pinecone, Weaviate, Chroma, and pgvector (a PostgreSQL extension).
Q85.What is an AI agent?
An AI agent is an AI system that can take actions in the world, not just generate text. Agents are given tools such as web search, code execution, file access, or API calls, and can autonomously plan and execute sequences of actions to accomplish a goal. Agents operate in a loop: they observe, reason, act, and observe again until a task is complete.
Q86.What is agentic AI?
Agentic AI refers to AI systems that operate with greater autonomy, taking multi-step actions, making decisions, and using tools over extended periods with limited human oversight. Agentic systems differ from single-turn chatbots in that they can pursue goals across many steps. Agentic AI introduces new challenges around reliability, safety, and oversight.
Q87.What is multi-agent AI?
Multi-agent AI systems use multiple AI models that collaborate or specialize to solve complex problems. One model might plan tasks, another executes them, and another reviews outputs. Multi-agent frameworks such as AutoGen and CrewAI enable more sophisticated automation by decomposing complex workflows across specialized agents.
Q88.What is function calling in LLMs?
Function calling (also called tool use) is a capability that allows LLMs to trigger external functions or APIs in a structured way. Instead of generating plain text, the model outputs a structured request to call a specified function with given parameters. This allows LLMs to perform actions such as retrieving data, running calculations, sending messages, or updating records in external systems.
Q89.How do you evaluate LLM output quality?
LLM output quality can be evaluated using automated metrics, human evaluation, and task-specific benchmarks. For production systems, creating a test set of representative inputs with expected outputs and regularly running evaluations against it is best practice. Human evaluation remains the gold standard for open-ended generation tasks, as automated metrics often miss nuance.
Q90.What is AI orchestration?
AI orchestration refers to the coordination of multiple AI models, tools, and data sources to complete complex tasks. Orchestration frameworks like LangChain and LlamaIndex provide tooling for building pipelines that sequence prompts, retrieve documents, call external APIs, and route between different models. Orchestration is the foundation of sophisticated AI agent systems.
10The Future of AI
Q91.What is AGI (Artificial General Intelligence)?
Artificial General Intelligence (AGI) refers to a hypothetical AI system that can perform any intellectual task that a human can, with comparable or superior capability. AGI would represent a qualitative shift from today's narrow AI systems, which excel at specific tasks but lack general-purpose intelligence. There is significant debate among researchers about the definition, timeline, and feasibility of AGI.
Q92.What is the difference between narrow AI and AGI?
Narrow AI (also called weak AI) refers to AI systems designed for specific tasks such as image recognition, language translation, or playing chess. All commercially deployed AI systems today are narrow AI. AGI would be an AI system capable of performing any cognitive task a human can, without needing to be specifically trained for each new task.
Q93.What are AI scaling laws?
AI scaling laws describe predictable relationships between the amount of compute, data, and model size used in training and the resulting performance of AI models. Research from OpenAI, DeepMind, and others has shown that increasing these factors leads to consistent, measurable improvements. Scaling laws have driven much of the rapid capability growth in AI over recent years.
Scaling Laws paper (arXiv)Q94.What are AI reasoning models?
AI reasoning models are LLMs specifically optimized for complex, multi-step reasoning tasks such as mathematics, coding, and logical problem solving. Examples include OpenAI's o1 and o3 series, which are trained to "think" through problems with extended internal reasoning before producing a final answer. Reasoning models trade speed and cost for significantly improved accuracy on hard problems.
Q95.What are AI coding assistants?
AI coding assistants are tools that integrate LLMs into the software development workflow to help developers write, review, debug, and document code. Examples include GitHub Copilot, Cursor, and Claude Code. These tools can autocomplete code, generate functions from natural language descriptions, explain existing code, fix bugs, and write tests, significantly accelerating development productivity.
Q96.What is the future of AI in the workplace?
AI is expected to augment most knowledge-work roles by automating routine and repetitive tasks, freeing workers to focus on higher-judgment activities. Research from McKinsey, Goldman Sachs, and the World Economic Forum suggests that AI will transform rather than simply eliminate most job categories. New roles related to AI oversight, prompt engineering, and AI integration are emerging rapidly.
McKinsey: Generative AI and the future of workQ97.Will AI replace jobs?
AI will automate specific tasks within jobs rather than replacing entire roles wholesale in most cases. Jobs involving routine, predictable, and text-based work face the highest exposure to automation. Jobs requiring physical dexterity, social and emotional intelligence, creative judgment, and novel problem-solving are less susceptible. Most economic forecasts expect AI to create new categories of work even as it displaces others.
Q98.What is AI regulation?
AI regulation refers to legal and policy frameworks that govern the development and deployment of AI systems. Regulatory approaches vary widely: the EU AI Act is a comprehensive risk-based framework; the United States has issued executive orders and sector-specific guidance; China has enacted regulations on generative AI content. Regulation is evolving rapidly alongside AI capabilities.
Q99.What is open-source AI?
Open-source AI refers to AI models, datasets, and tools whose weights and code are publicly released, allowing anyone to inspect, use, and modify them. Organizations like Meta (Llama), Mistral AI, and Stability AI have released open models. The Open Source Initiative has published a formal definition of open-source AI to clarify what "open" means in the AI context.
Open Source AI DefinitionQ100.What is the current state of AI development?
As of 2025, AI development is advancing rapidly, with frontier models demonstrating strong performance across language, reasoning, coding, and multimodal tasks. The field is characterized by intensive competition among major AI labs, growing regulatory attention, and rapid commercial deployment across industries. Key open challenges include reducing hallucinations, improving reliability in agentic systems, and managing the societal implications of accelerating AI capabilities.
Stanford AI Index Report