How to Learn the Basics of AI and LLMs (Without a PhD)
You don’t need a PhD to understand or use AI and large language models. In a few focused weeks you can go from “what even is ChatGPT?” to “I built something that uses an LLM”—if you learn the right 20% and skip the rest.
This guide is for you if: you’re curious how tools like ChatGPT and Claude work, you want to use AI in a product or side project, or you need to have sensible conversations about AI at work—without diving into research papers.
Here’s what LLMs actually are, what to learn first, what to skip, and how to get to “I can build with this” fast.
What “the basics” means (and what it doesn’t)
You won’t be training models from scratch—and you don’t need to. For most people, “basics” means:
- What an LLM is — in plain language: a model that predicts the next token given context. Tokens are the atomic units (words, subwords, or characters); for English, roughly one token ≈ 4 characters or about ¾ of a word. That’s why it can chat, summarize, and follow instructions.
- Prompts and prompt engineering — how to ask for what you want, iterate, and get usable output. This is 80% of day-to-day use.
- APIs and tools — calling an LLM from code (e.g. OpenAI, Anthropic, or open-source APIs) so you can build apps, not just chat in a UI.
- When LLMs work well (and when they don’t) — strengths (drafting, summarization, structured output) and limits (hallucination, context length, cost). So you don’t over- or under-use them.
That’s enough to use AI in a side project, evaluate tools at work, or understand what “AI-built” features actually do. It’s a realistic goal without a math or CS degree.
What to skip at first (so you don’t get lost)
Stay focused so you actually ship:
- Deep ML theory — transformers, attention, training. Interesting later; not required to use LLMs.
- Fine-tuning and training — building or retraining models. Most people only need to call existing APIs and write good prompts.
- Every model and vendor — pick one API (e.g. OpenAI or Anthropic), learn it well, then add others if you need them.
- Hype and fear cycles — focus on what’s actually usable today: text in, text (or structured data) out.
Nail concepts + prompting + one API first. The rest is incremental.
What to learn first (in order)
Step 1: What an LLM is (about 1–2 hours)
- Plain-language mental model: An LLM estimates the probability of token sequences—it takes text (your prompt + prior messages) and predicts what comes next. It’s trained on huge amounts of text, so it can mimic patterns: conversation, code, formats. That enables text generation, translation, summarization, and more.
- Key terms: prompt, completion, token (the atomic unit; in English ~4 characters or ~¾ of a word), context window (how much text the model can “see” at once), temperature (how random vs deterministic the output is).
- Why it’s useful: You can get draft text, summaries, classifications, structured data (JSON), and simple reasoning—all from a well-written prompt.
You don’t need to implement anything here. Free short courses (e.g. Coursera’s “Introduction to Large Language Models,” ~1 hour) can give you this overview in an afternoon. Just enough so “I send text in, I get text back” makes sense.
Step 2: Prompting that works (about 1 week of practice)
Prompt engineering is the process of writing effective instructions so models consistently produce the output you want. It’s often described as both an art and a science—non-deterministic, but with techniques that help. Key approaches (as reflected in OpenAI and other docs, 2024):
- Instruction prompting — clear, concise instructions (e.g. “Summarize this in 3 bullet points for a busy exec” beats “Summarize this”).
- Role prompting — assign a role to control tone and style (“You are a technical writer”).
- Few-shot prompting — include 1–2 input/output examples in the prompt so the model mimics the format. Zero-shot (no examples) and one-shot (one example) are also common.
- Give structure — “Return a JSON object with keys: title, summary, action_items” so you can use the output in code.
- Iterate — first prompt is rarely perfect. Refine based on what you get; this is the main skill.
- Guardrails — don’t assume the output is always correct. Check critical facts; use structured output when you need to parse it.
Spend a few days in ChatGPT or Claude: summarization, extraction, simple reasoning, code generation. A free resource like Learn Prompting offers a beginner’s guide to generative AI and prompt-writing. Notice what works and what doesn’t—that’s prompt engineering.
Step 3: One API from code (about 1 week)
- Pick one provider — e.g. OpenAI (e.g.
gpt-4,gpt-3.5-turbo) or Anthropic (e.g.claude-3-sonnet). Check each provider’s latest model list and docs. OpenAI’s Developer quickstart covers creating an API key, storing it in an env var (e.g.OPENAI_API_KEY), and using the SDK (Python, JavaScript, etc.). For chat-style apps, you’ll use a “chat completions” or “responses” endpoint: send a list of messages (system + user, or user + assistant + user) and get back the model’s reply. - Basic flow: Send messages; get back a message with the model’s reply. Parse it and use it in your app. Consider trade-offs: speed vs cost vs capability (e.g. reasoning models are more capable but slower and more expensive).
- Environment and keys — store API keys in env vars, never in code. Use a small script or a minimal web app (e.g. Next.js API route, or a Python script) that calls the API and prints or returns the result.
- Structured output — many APIs support JSON mode or a response format. Use it when you need to drive logic (e.g. “extract these fields and save to DB”). For production, pin to a specific model version and consider building a simple way to monitor prompt performance.
By the end of this you should be able to: “From my app, send a prompt and get back text or JSON I can use.” That’s the foundation for any LLM-powered feature.
Step 4: One small project (about 1–2 weeks)
Build something you’d actually use or show:
- A small chatbot (custom system prompt, chat history in context).
- A summarizer or categorizer (paste text → get summary or tags).
- A simple “AI agent” that uses tools (e.g. search + LLM) in a loop.
Use your chosen API, keep scope small, and ship. You’ll learn more from one shipped project than from five tutorials.
How long it usually takes
Evidence from 2024–2025 suggests:
- “I get how this works” — You can grasp core LLM concepts in a few days (or even ~1 hour with a short intro course). No technical background required for concept-level courses (e.g. Coursera’s intro to LLMs, Learn Prompting’s intro).
- “I can write a good prompt and use the UI” — 1–2 weeks: concepts + prompting in ChatGPT, Claude, or similar.
- “I can call the API from code” — 2–3 weeks: add one API and a tiny script or endpoint. Focused “Introduction to OpenAI”–style courses are in the ~9 hour range; you can spread that over 2–3 weeks.
- “I built one small thing that uses an LLM” — 4–6 weeks at a few hours per week: one project (chatbot, summarizer, or simple agent). Some “practical AI skills” programs report ~6 weeks at ~4 hours/week for hands-on proficiency with AI tools.
If you have more time per week, you compress this. The bottleneck is usually practice, not theory.
How to avoid tutorial hell
- Use the product first — ChatGPT, Claude, or similar. Build intuition before you write code.
- One API, one language — don’t learn OpenAI + Anthropic + LangChain in week 1. Nail one stack.
- Build one thing — even if it’s tiny. “Summarize my clipboard” or “FAQ bot from a doc” is enough to make it concrete.
- Ignore the hype — you don’t need to know every new model. Understand the basics; add new tools when you have a real need.
If you’d rather skip the resource hunt and get a path built for your goal and schedule, you can get a custom course in minutes →. Describe what you want to learn (e.g. “AI and LLM basics for a side project, 2 hours a week”) and get a structured plan—nothing you don’t need.
Bottom line
You can learn the basics of AI and LLMs without a PhD by focusing on: (1) a simple mental model of what an LLM is, (2) prompting that’s specific and structured, (3) one API from code, and (4) one small project. Skip deep ML theory and training at first; get to “I understand it and I can build with it,” then go deeper if you want.
Skip the tutorial hunt. Tell us what you want to learn and how much time you have (e.g. “LLM basics to build a small AI feature, 3 hours a week”), and we’ll build you a custom course—structured lessons, in the right order. Build my course →