LogoAIAny
  • Search
  • Collection
  • Category
  • Tag
  • Blog
LogoAIAny

Tag

Explore by tags

LogoAIAny

Learn Anything about AI in one site.

support@aiany.app
Product
  • Search
  • Collection
  • Category
  • Tag
Resources
  • Blog
Company
  • Privacy Policy
  • Terms of Service
  • Sitemap
Copyright © 2025 All Rights Reserved.
  • All

  • 30u30

  • ASR

  • ChatGPT

  • GNN

  • IDE

  • RAG

  • ai-agent

  • ai-api

  • ai-api-management

  • ai-client

  • ai-coding

  • ai-demos

  • ai-development

  • ai-framework

  • ai-image

  • ai-image-demos

  • ai-inference

  • ai-leaderboard

  • ai-library

  • ai-rank

  • ai-serving

  • ai-tools

  • ai-train

  • ai-video

  • ai-workflow

  • AIGC

  • alibaba

  • amazon

  • anthropic

  • audio

  • blog

  • book

  • bytedance

  • chatbot

  • chemistry

  • claude

  • course

  • deepmind

  • deepseek

  • engineering

  • foundation

  • foundation-model

  • gemini

  • google

  • gradient-booting

  • grok

  • huggingface

  • LLM

  • math

  • mcp

  • mcp-client

  • mcp-server

  • meta-ai

  • microsoft

  • mlops

  • NLP

  • nvidia

  • openai

  • paper

  • physics

  • plugin

  • RL

  • science

  • sora

  • translation

  • tutorial

  • vibe-coding

  • video

  • vision

  • xAI

  • xai

Icon for item

nanochat

2025
Andrej Karpathy

nanochat is a full-stack, minimal codebase for training, fine-tuning, evaluating, and deploying a ChatGPT-like large language model (LLM) from scratch on a single 8xH100 GPU node for under $100.

LLMchatbotai-trainai-toolstutorial+1
Icon for item

SWE-agent: Agent-Computer Interfaces Enable Automated Software Engineering

2024
John Yang, Carlos E. Jimenez +5

SWE-agent is a system designed to empower language model (LM) agents to autonomously perform software engineering tasks. It features a custom agent-computer interface (ACI) that enhances the agent's ability to navigate repositories, create and edit code, and execute programs, achieving state-of-the-art results on the SWE-bench and HumanEvalFix benchmarks. [2, 5, 8]

paperai-agentLLMai-codingengineering
Icon for item

OpenAI Agents SDK

2025
OpenAI

The OpenAI Agents SDK is a lightweight yet powerful framework for building multi-agent workflows. It is provider-agnostic, supporting the OpenAI Responses and Chat Completions APIs, as well as 100+ other LLMs via integrations like LiteLLM.

openaiai-agentai-frameworkai-developmentai-library+1
Icon for item

Agent Lightning

2025
Microsoft Research

Agent Lightning is an open-source framework developed by Microsoft Research for optimizing and training AI agents using reinforcement learning (RL) and other techniques, supporting integration with any agent framework with minimal code changes.

RLLLMai-agentmicrosoftai-train+3
Icon for item

ReAct: Synergizing Reasoning and Acting in Language Models

2022
Shunyu Yao, Jeffrey Zhao +5

This paper introduces ReAct, an approach that integrates reasoning and acting in large language models (LLMs). ReAct enables LLMs to generate both reasoning traces and task-specific actions in an interleaved manner. This synergy allows reasoning to help induce, track, and update action plans, while actions interface with external sources like knowledge bases to gather more information, overcoming issues of hallucination and error propagation in prior methods.

paperLLMNLPai-agentgoogle+1
Icon for item

Unsloth

2023
Daniel Han, Michael Han

Open-source framework that accelerates fine-tuning and full training of transformer LLMs by up to 30 × while cutting VRAM requirements by roughly 90 %, letting developers train custom models quickly on commodity GPUs.

ai-developmentai-libraryai-trainLLM
Icon for item

Ollama

2023
Jeffrey Morgan, Michael Chiang

A lightweight open-source platform for running, managing, and integrating large language models locally via a simple CLI and REST API.

ai-developmentai-libraryai-inferenceai-servingLLM
Icon for item

NVIDIA NeMo

2019
NVIDIA

End-to-end NVIDIA framework and micro-services platform for building, customizing, and deploying large language, speech, vision, and multimodal AI models.

ai-developmentai-libraryai-trainLLMnvidia

Attention Is All You Need

2017
Ashish Vaswani, Noam Shazeer +6

The paper “Attention Is All You Need” (2017) introduced the Transformer — a novel neural architecture relying solely on self-attention, removing recurrence and convolutions. It revolutionized machine translation by dramatically improving training speed and translation quality (e.g., achieving 28.4 BLEU on English-German tasks), setting new state-of-the-art benchmarks. Its modular, parallelizable design opened the door to large-scale pretraining and fine-tuning, ultimately laying the foundation for modern large language models like BERT and GPT. This paper reshaped the landscape of NLP and deep learning, making attention-based models the dominant paradigm across many tasks.

NLPLLMAIGC30u30paper+1

Relational recurrent neural networks

2018
Adam Santoro, Ryan Faulkner +8

This paper introduces a Relational Memory Core that embeds multi-head dot-product attention into recurrent memory to enable explicit relational reasoning. Evaluated on synthetic distance-sorting, program execution, partially-observable reinforcement learning and large-scale language-modeling benchmarks, it consistently outperforms LSTM and memory-augmented baselines, setting state-of-the-art results on WikiText-103, Project Gutenberg and GigaWord. By letting memories interact rather than merely store information, the approach substantially boosts sequential relational reasoning and downstream task performance.

foundation30u30paperNLPLLM

GPT2: Language Models are Unsupervised Multitask Learners

2019
Alec Radford, Jeffrey Wu +4

This paper introduces GPT-2, showing that large-scale language models trained on diverse internet text can perform a wide range of natural language tasks in a zero-shot setting — without any task-specific training. By scaling up to 1.5 billion parameters and training on WebText, GPT-2 achieves state-of-the-art or competitive results on benchmarks like language modeling, reading comprehension, and question answering. Its impact has been profound, pioneering the trend toward general-purpose, unsupervised language models and paving the way for today’s foundation models in AI.

LLMNLPopenaipaper
Icon for item

n8n

2019
n8n GmbH

Open-source, node-based workflow-automation platform for designing and running complex integrations and AI-powered flows.

LLMai-libraryai-developmentai-frameworkai-agent+1
  • Previous
  • 1
  • 2
  • 3
  • Next