Skip to content
No results
Starphix
  • Catalog
  • Roadmap Builder
  • StarphiX HQ
    • About
    • Haive
    • PaiX
    • Policy
    • Terms
    • Jobs
Shopping cart$0.00 0
Learn AI
Starphix
  • Catalog
  • Roadmap Builder
  • StarphiX HQ
    • About
    • Haive
    • PaiX
    • Policy
    • Terms
    • Jobs
Shopping cart$0.00 0
Learn AI
Starphix

Welcome | Guided Learning Paths

  • Welcome to the StarphiX Knowledge Center!
  • 🧭 Curated Learning Paths
    • The Learning Path for the Student & Creative 🎨
    • The Learning Path for the Developer & Tech Enthusiast 💻
    • The Learning Path for the Business Owner & Professional 💼

The Story of AI: Past, Present, & Future

  • Pillar I: 📖
  • 📜 A Brief History of AI
    • The Transformer Revolution: The Architecture That Changed Everything 🧠
    • The Rise of Machine Learning: A New Paradigm 📈
    • The AI Winters: When Promises Outpaced Reality ❄️
    • The Dartmouth Workshop: The Birth of a Field 💡
    • The Dream of an Artificial Mind: AI’s Philosophical Origins 🏛️
  • 🌍 The AI Landscape Today
    • An Overview of AI’s Impact on Modern Work & Creativity 💼
    • Generative AI vs. Traditional AI: What’s the Difference? ↔️
    • Why Now? Understanding the Current AI Boom 💥
  • 🔭 The Future of AI: The Next Frontier
    • An Introduction to AI Ethics & Responsible Development ⚖️
    • An Introduction to AI Ethics & Responsible Development ⚖️
    • AI for Good: The Role of AI in Science, Medicine, and Climate Change ❤️
    • The Quest for AGI: What is Artificial General Intelligence? 🤖

The Modern AI Toolkit

  • ⚙️ The Technology Stack Explained
    • The Hardware Layer: Why GPUs are the Engine of AI ⚙️
    • The Model Layer: Understanding LLMs, Diffusion Models, and Agents 🧠
    • The Platform Layer: How APIs and No-Code Tools Connect Everything 🔗
  • 🏢 The Ecosystem: Major Players & Platforms
    • Major Players & Platforms 🏢
  • 🛠️ Practical Use Cases by Profession
    • For the Small Business Owner: 5 High-Impact Automations to Implement Today 🧑‍💼
    • For the Consultant or Coach: Streamlining Your Client Workflow with AI 🧑‍🏫
    • For the Creative Professional: Using AI as a Brainstorming Partner, Not a Replacement 🎨
    • For the Student & Researcher: How to Supercharge Your Learning with AI 🧑‍🎓

The Sovereign AI: A Guide to Local Systems

  • 🧠 The Philosophy of AI Sovereignty
    • Why Local AI is the Future of Work and Creativity 🚀
    • Data Privacy vs. Data Sovereignty: Taking Control of Your Digital Self 🛡️
    • The Open-Source AI Movement: A Force for Democratization 🌐
  • 🏠 Your First Local AI Lab
    • Understanding the Core Components of a Local AI Setup 🖥️
    • Choosing Your Hardware: A Buyer’s Guide for Every Budget 💰
    • The Software Stack: A Step-by-Step Installation Guide 💿
    • Downloading Your First Open-Source Model 🧠
    • A Guide to Model Sizes: What Do 7B, 13B, and 70B Really Mean? 📏
  • 🏗️ Building with Local AI: Practical Workflows
    • Your First Local Automation: Connecting to n8n 🤖
    • Creating a Private Chat Interface for Your Local Models 💬
    • The Power of APIs: Connecting Local AI to Other Tools 🔗
    • Practical Project: Building a Private ‘Meeting Matrix Summarizer’ 📄
    • Practical Project: Creating a ‘Knowledge-Core Agent’ with Your Own Documents 🧠
  • 🚀 Advanced Concepts & The PaiX Vision
    • An Introduction to Fine-Tuning Your Own Models ⚙️
    • Optimizing Performance: Quantization and Model Pruning Explained ⚡️
    • The StarphiX Vision: From DIY Homelab to a Professional PaiX Local Workstation ✨

The Library: Resources & Reference

  • The Archive of Seminal Papers 📜
  • Glossary of AI Terms 📖
  • The Directory of Tools & Frameworks 🧰
View Categories
  • Home
  • Docs
  • The Library: Resources & Reference
  • The Archive of Seminal Papers 📜

The Archive of Seminal Papers 📜

4 min read

Introduction #

Welcome to the Archive. This is a curated collection of the foundational academic papers that have defined and shaped the field of modern Artificial Intelligence. While the papers themselves can be dense and technical, they represent the source code of the ideas that power our world today. For each entry, we’ve provided a direct link to the paper and a simple “Why This Paper Matters” summary to explain its core contribution and historical impact in plain language.


Computing Machinery and Intelligence (1950) #

  • Authors: Alan M. Turing
  • Publication: Mind, LIX (236): 433–460
  • Link to Paper
  • Why This Paper Matters: This is the philosophical origin of modern AI. In this paper, Turing asked the question, “Can machines think?” and proposed the famous “Turing Test” as a way to measure a machine’s ability to exhibit intelligent behavior indistinguishable from that of a human. It set the stage for the entire field.

ImageNet Classification with Deep Convolutional Neural Networks (2012) #

  • Authors: Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton
  • Publication: Advances in Neural Information Processing Systems 25 (NIPS 2012)
  • Link to Paper
  • Why This Paper Matters: This paper, which introduced the “AlexNet” architecture, is widely seen as the “Big Bang” moment for the modern deep learning boom. By winning the ImageNet competition by a massive margin, it proved that deep neural networks, combined with powerful GPUs, could solve complex problems at a scale previously thought impossible.

Attention Is All You Need (2017) #

  • Authors: Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, Illia Polosukhin
  • Publication: Advances in Neural Information Processing Systems 30 (NIPS 2017)
  • Link to Paper
  • Why This Paper Matters: This is arguably the most important AI paper of the last decade. It introduced the Transformer architecture, which solved a key memory problem in older models and allowed for massive parallel processing of language. Every modern Large Language Model, including ChatGPT and Gemini, is a direct descendant of the ideas in this paper.

Language Models are Unsupervised Multitask Learners (2019) #

  • Authors: Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever
  • Publication: OpenAI
  • Link to Paper
  • Why This Paper Matters: This paper introduced GPT-2, a model so powerful for its time that OpenAI initially released it in stages due to concerns about misuse. It proved that by dramatically scaling up a Transformer model and training it on a massive, diverse dataset, a single model could perform a wide range of language tasks without explicit training for each one.

Language Models are Few-Shot Learners (2020) #

  • Authors: Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, et al.
  • Publication: Advances in Neural Information Processing Systems 33 (NeurIPS 2020)
  • Link to Paper
  • Why This Paper Matters: This is the GPT-3 paper. It demonstrated the incredible power of “scaling laws,” showing that as models get even bigger, new abilities emerge. GPT-3 could perform tasks with only a few examples (“few-shot learning”), setting the standard for large language models and leading directly to the creation of ChatGPT.

Denoising Diffusion Probabilistic Models (2020) #

  • Authors: Jonathan Ho, Ajay Jain, Pieter Abbeel
  • Publication: Advances in Neural Information Processing Systems 33 (NeurIPS 2020)
  • Link to Paper
  • Why This Paper Matters: This paper was a breakthrough for image generation. It introduced a refined and highly effective approach to diffusion models, which create images by starting with random noise and gradually refining it into a coherent picture. This work laid the foundation for modern text-to-image systems like DALL-E 2 and Stable Diffusion.

Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks (2020) #

  • Authors: Patrick Lewis, Ethan Perez, Aleksandara Piktus, et al.
  • Publication: Advances in Neural Information Processing Systems 33 (NeurIPS 2020)
  • Link to Paper
  • Why This Paper Matters: This paper formally introduced RAG, the powerful technique for making AI models more accurate and reliable. By teaching a model to first retrieve relevant information from a knowledge base before generating an answer, RAG helps reduce “hallucinations” and allows AI to use up-to-date or private information. It is the core technology behind our “Knowledge-Core Agent” project.

LoRA: Low-Rank Adaptation of Large Language Models (2021) #

  • Authors: Edward J. Hu, Yelong Shen, Phillip Wallis, et al.
  • Publication: International Conference on Learning Representations (ICLR 2022)
  • Link to Paper
  • Why This Paper Matters: This paper made advanced customization of AI accessible to a much wider audience. It introduced LoRA, a “parameter-efficient” fine-tuning method that allows developers to specialize large models for new tasks using a tiny fraction of the computational power of a full fine-tune. This breakthrough is critical for the local and open-source AI communities.
Table of Contents
  • Introduction
    • Computing Machinery and Intelligence (1950)
    • ImageNet Classification with Deep Convolutional Neural Networks (2012)
    • Attention Is All You Need (2017)
    • Language Models are Unsupervised Multitask Learners (2019)
    • Language Models are Few-Shot Learners (2020)
    • Denoising Diffusion Probabilistic Models (2020)
    • Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks (2020)
    • LoRA: Low-Rank Adaptation of Large Language Models (2021)
  • About
  • Policy
  • Terms
  • Jobs
  • StarphiX HQ

Copyright © 2025 | PaiX Built