LogoAIAny
  • Search
  • Collection
  • Category
  • Tag
  • Blog
LogoAIAny

Learn Anything about AI in one site

Best learning resources for AI

LogoAIAny

Learn Anything about AI in one site.

support@aiany.app
Product
  • Search
  • Collection
  • Category
  • Tag
Resources
  • Blog
Company
  • Privacy Policy
  • Terms of Service
  • Sitemap
Copyright © 2025 All Rights Reserved.
Icon for item

OpenVINO

2018
Intel

OpenVINO is an open-source toolkit from Intel that streamlines the optimization and deployment of AI inference models across a wide range of Intel® hardware.

ai-developmentai-inferenceai-serving

GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism

2018
Yanping Huang, Youlong Cheng +9

This paper introduces GPipe, a model-parallelism library designed to train large neural networks efficiently using pipeline parallelism. It partitions models across accelerators, processes micro-batches in parallel, and supports synchronous gradient updates. GPipe enables near-linear scaling with the number of devices while maintaining model quality and training stability. It achieves state-of-the-art performance in large-scale image classification (AmoebaNet) and multilingual machine translation (6B parameter Transformer), demonstrating flexibility across tasks. Its impact lies in making massive model training more practical and accessible across diverse architectures without relying on high-speed interconnects or custom model designs.

foundation30u30paperengineering
Icon for item

ONNX Runtime

2018
Microsoft

Microsoft’s high-performance, cross-platform inference engine for ONNX and GenAI models.

ai-developmentai-inferenceai-serving
Icon for item

MLRun

2019
Iguazio / QuantumBlack

An Iguazio-backed open-source framework that orchestrates data/ML/LLM pipelines with serverless execution, tracking and monitoring.

ai-developmentmlops

GPT2: Language Models are Unsupervised Multitask Learners

2019
Alec Radford, Jeffrey Wu +4

This paper introduces GPT-2, showing that large-scale language models trained on diverse internet text can perform a wide range of natural language tasks in a zero-shot setting — without any task-specific training. By scaling up to 1.5 billion parameters and training on WebText, GPT-2 achieves state-of-the-art or competitive results on benchmarks like language modeling, reading comprehension, and question answering. Its impact has been profound, pioneering the trend toward general-purpose, unsupervised language models and paving the way for today’s foundation models in AI.

LLMNLPopenaipaper
Icon for item

n8n

2019
n8n GmbH

Open-source, node-based workflow-automation platform for designing and running complex integrations and AI-powered flows.

LLMai-libraryai-developmentai-frameworkai-agent+1
Icon for item

Megatron-LM

2019
NVIDIA

NVIDIA’s model-parallel training library for GPT-like transformers at multi-billion-parameter scale.

ai-developmentai-frameworkai-trainnvidia
Icon for item

BentoML

2019
BentoML Team

Open-source framework for building, shipping and running containerized AI services with a single command.

ai-developmentai-inferenceai-serving
Icon for item

Metaflow

2019
Netflix

Netflix’s human-centric framework for building and operating real-life data-science and ML workflows with idiomatic Python and production-grade scaling.

ai-developmentmlops
Icon for item

Flyte

2020
Lyft, Flyte Community

A Kubernetes-native workflow engine (originally at Lyft, now LF AI & Data) that provides strongly-typed, versioned data/ML pipelines at scale.

ai-developmentmlops

Scaling Laws for Neural Language Models

2020
Jared Kaplan, Sam McCandlish +8

reveals that language model performance improves predictably as you scale up model size, dataset size, and compute, following smooth power-law relationships. It shows that larger models are more sample-efficient, and optimally efficient training uses very large models on moderate data, stopping well before convergence. The work provided foundational insights that influenced the development of massive models like GPT-3 and beyond, shaping how the AI community understands trade-offs between size, data, and compute in building ever-stronger models.

LLMNLPopenai30u30paper

GPT3: Language Models are Few-Shot Learners

2020
Tom B. Brown, Benjamin Mann +29

This paper introduces GPT-3, a 175-billion-parameter autoregressive language model that achieves impressive zero-shot, one-shot, and few-shot performance across diverse NLP tasks without task-specific fine-tuning. Its scale allows it to generalize from natural language prompts, rivaling or surpassing prior state-of-the-art models that require fine-tuning. The paper’s impact is profound: it demonstrated the power of scaling laws, reshaped research on few-shot learning, and sparked widespread adoption of large-scale language models, influencing advancements in AI applications, ethical debates, and commercial deployments globally.

LLMNLPopenaipaper
  • Previous
  • 1
  • More pages
  • 8
  • 9
  • 10
  • More pages
  • 17
  • Next