LlamaIndex
Skill Verified ActiveData framework for building LLM applications with RAG. Specializes in document ingestion (300+ connectors), indexing, and querying. Features vector indices, query engines, agents, and multi-modal support. Use for document Q&A, chatbots, knowledge retrieval, or building RAG pipelines. Best for data-centric LLM applications.
To enable users to build robust LLM applications that leverage their own data through Retrieval-Augmented Generation, offering specialized tools for document ingestion, indexing, and querying.
Features
- Document ingestion from 300+ connectors
- Data indexing and querying capabilities
- Support for vector indices
- Integration with query engines and agents
- Multi-modal RAG support
Use Cases
- Building RAG applications
- Document question-answering over private data
- Creating knowledge bases for LLMs
- Building chatbots with enterprise data
Non-Goals
- Serving as a general-purpose agent framework without a RAG focus
- Replacing a vector database directly (integrates with them)
- Providing a UI for end-users (focuses on developer framework)
Practices
- RAG implementation
- Data ingestion
- Vector indexing
- LLM application development
Prerequisites
- Python 3.8+
- pip package manager
- API keys for LLMs (e.g., OpenAI, Anthropic) if used
Installation
First, add the marketplace
/plugin marketplace add Orchestra-Research/AI-Research-SKILLs/plugin install AI-Research-SKILLs@ai-research-skillsQuality Score
VerifiedTrust Signals
Similar Extensions
LangChain
99Framework for building LLM-powered applications with agents, chains, and RAG. Supports multiple providers (OpenAI, Anthropic, Google), 500+ integrations, ReAct agents, tool calling, memory management, and vector store retrieval. Use for building chatbots, question-answering systems, autonomous agents, or RAG applications. Best for rapid prototyping and production deployments.
Embedding Strategies
100Select and optimize embedding models for semantic search and RAG applications. Use when choosing embedding models, implementing chunking strategies, or optimizing embedding quality for specific domains.
Dspy
99Build complex AI systems with declarative programming, optimize prompts automatically, create modular RAG systems and agents with DSPy - Stanford NLP's framework for systematic LM programming
Langchain Framework
99LangChain LLM application framework with chains, agents, RAG, and memory for building AI-powered applications
Hybrid Search Implementation
98Combine vector and keyword search for improved retrieval. Use when implementing RAG systems, building search engines, or when neither approach alone provides sufficient recall.
Init
100Creates, updates, or optimizes an AGENTS.md file for a repository with minimal, high-signal instructions covering non-discoverable coding conventions, tooling quirks, workflow preferences, and project-specific rules that agents cannot infer from reading the codebase. Use when setting up agent instructions or Claude configuration for a new repository, when an existing AGENTS.md is too long, generic, or stale, when agents repeatedly make avoidable mistakes, or when repository workflows have changed and the agent configuration needs pruning. Applies a discoverability filter—omitting anything Claude can learn from README, code, config, or directory structure—and a quality gate to verify each line remains accurate and operationally significant.