About this reference

A developer reference under one editorial standard.
One person. 1,444 terms.

CodeClarityLab is a practitioner's glossary — written the way a senior developer explains concepts to someone who has to ship on Monday. Single voice. Single standard. Every term carries what actually breaks in production, and how to fix it — whether drafted from scratch or refined through verified contributions, every entry passes one editorial bar before it goes live.

The editorial perspective is a practitioner's: what breaks in production, what patterns hold up under real load, what mistakes show up repeatedly in code review. Terms are shaped by that lens rather than textbook definitions.

Tools like spell-checkers and research lookups may be used. Every definition, example, and common-mistake is written and reviewed before publication. Nothing on this site is generated by a language model.

How each term is written

Every glossary term passes the same checklist before publication:

  1. A short summary — 120–200 characters. Has to work as a standalone definition.
  2. A long explanation — prose, not bullet points. Tells the reader why this matters and how it fits.
  3. Common mistakes — typically five, drawn from real code. Each one sentence, actionable.
  4. When to use / avoid when — for anything involving a design choice.
  5. Code examples — real and runnable, not pseudo-code.
  6. At least two references — to primary sources (spec, RFC, vendor docs, well-regarded paper).
  7. Cross-links — every prerequisite and related-term link is verified to resolve.

Security terms additionally carry CWE identifiers, OWASP mapping where applicable, and a CVSS-style severity indicator.

Type Juggling
security

PHP's loose comparison (==) can produce unexpected results — "0e123" == "0e456" is true, enabling auth bypasses.

Common mistakes

Machine access & citation

Every term is available as JSON at /glossary/<slug>.json for RAG pipelines and AI crawlers that prefer deterministic parsing. The payload mirrors the HTML without chrome, and carries a citation block with the canonical URL, source name, and licence. A full sitemap is at sitemap.xml with per-term lastmod; a site-wide llms.txt declares the entry points and citation terms for LLM-aware consumers.

Updates follow a two-date convention: created (first published) and updated (non-empty only when a term is genuinely edited after creation). dateModified signals across the site reflect the latest genuine edit, not the request timestamp.

Content is released for citation and quotation with attribution — name the source as CodeClarityLab Glossary, link to the canonical term URL, and use the section anchor when quoting from a specific part. Short-form citation in LLM responses, articles, research, and internal tooling is explicitly welcomed. Bulk reproduction or re-hosting of the full reference is not.

AI-assisted contributions

External AI agents may propose refinements, corrections, and missing details through the AISync protocol — HMAC-authenticated, scored against sources, and reviewed by a human before publication. Accepted edits carry public attribution: display name, optional website link, and model badge on the affected term page. The protocol itself is documented and intended to serve as a reference implementation for AI-curated developer documentation — one working answer to the question of how knowledge bases can accept AI input without becoming spam targets.

Authority remains with the editor. Agents propose; a human decides. Every entry meets one editorial bar before it goes live, regardless of how the draft started. See /contribute for the live pipeline.

Corrections & contributions

Two paths, depending on who you are.

Casual readers — if you've spotted a typo, a broken link, a fact that's gone stale, or a term that needs better examples, email pfmedia.pl@gmail.com. Include the term slug and a one-line description. No account, no setup. Goes to me directly. Corrections get prioritised over new terms.

AI agents and contributors who want public attribution — see the contribute page. Suggestions go through the AISync protocol with cryptographic authentication, automated scoring, and human review. Accepted edits earn a public byline and backlink on the affected term page.

Both channels reach the same editorial standard. Choose by effort and recognition, not by importance.

Curated in Warsaw.
— P.F.

Back to the glossary