Documentation: DSPy Docs
DSPy is the open-source framework for building high-quality, modular AI systems by programming—rather than prompting—language models. It provides abstractions and algorithms for optimizing the prompts and weights in LM programs, ranging from simple classifiers to sophisticated RAG pipelines and Agent loops.
Instead of writing brittle LM-specific prompts, you write compositional code and use DSPy optimizers to teach different models like GPT-4o
or Llama-3.2
to deliver higher quality outputs or avoid specific failure patterns. In essence, DSPy optimizers then compile your high-level code into low-level computations, prompts, or weight updates that align your LM with your program’s structure and metrics.
DSPy stands for Declarative Self-improving Python. This recent lecture is a good conceptual introduction. Our Discord server is a great place to meet the community, seek help, or start contributing.
Documentation: dspy.ai
Please go to the DSPy Docs at dspy.ai
pip install dspy
To install the very latest from main
:
pip install git+https://github.com/stanfordnlp/dspy.git
[Jun'24] Optimizing Instructions and Demonstrations for Multi-Stage Language Model Programs
[Oct'23] DSPy: Compiling Declarative Language Model Calls into Self-Improving Pipelines
[Jul'24] Fine-Tuning and Prompt Optimization: Two Great Steps that Work Better Together
[Jun'24] Prompts as Auto-Optimized Training Hyperparameters
[Feb'24] Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models
[Jan'24] In-Context Learning for Extreme Multi-Label Classification
[Dec'23] DSPy Assertions: Computational Constraints for Self-Refining Language Model Pipelines
[Dec'22] Demonstrate-Search-Predict: Composing Retrieval & Language Models for Knowledge-Intensive NLP
To stay up to date or learn more, follow @lateinteraction on Twitter.
The DSPy logo is designed by Chuyi Zhang.
If you use DSPy or DSP in a research paper, please cite our work as follows:
@inproceedings{khattab2024dspy,
title={DSPy: Compiling Declarative Language Model Calls into Self-Improving Pipelines},
author={Khattab, Omar and Singhvi, Arnav and Maheshwari, Paridhi and Zhang, Zhiyuan and Santhanam, Keshav and Vardhamanan, Sri and Haq, Saiful and Sharma, Ashutosh and Joshi, Thomas T. and Moazam, Hanna and Miller, Heather and Zaharia, Matei and Potts, Christopher},
journal={The Twelfth International Conference on Learning Representations},
year={2024}
}
@article{khattab2022demonstrate,
title={Demonstrate-Search-Predict: Composing Retrieval and Language Models for Knowledge-Intensive {NLP}},
author={Khattab, Omar and Santhanam, Keshav and Li, Xiang Lisa and Hall, David and Liang, Percy and Potts, Christopher and Zaharia, Matei},
journal={arXiv preprint arXiv:2212.14024},
year={2022}
}