diff --git a/candle-transformers/src/models/bert.rs b/candle-transformers/src/models/bert.rs index bdc0385deb..a7db075cbb 100644 --- a/candle-transformers/src/models/bert.rs +++ b/candle-transformers/src/models/bert.rs @@ -1,3 +1,9 @@ +//! BERT (Bidirectional Encoder Representations from Transformers) +//! +//! See "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding", Devlin et al. 2018 +//! - [Arxiv](https://arxiv.org/abs/1810.04805) +//! - [Github](https://github.com/google-research/bert) +//! use super::with_tracing::{layer_norm, linear, LayerNorm, Linear}; use candle::{DType, Device, Result, Tensor}; use candle_nn::{embedding, Embedding, Module, VarBuilder}; diff --git a/candle-transformers/src/models/moondream.rs b/candle-transformers/src/models/moondream.rs index e714eb7f23..d351d7c019 100644 --- a/candle-transformers/src/models/moondream.rs +++ b/candle-transformers/src/models/moondream.rs @@ -1,4 +1,4 @@ -//! MoonDream Model provides vision-to-text capabilities based on Microsoft's Phi architecture +//! MoonDream Model vision-to-text //! //! The model consists of: //! - Vision encoder using a ViT-style architecture diff --git a/candle-transformers/src/models/persimmon.rs b/candle-transformers/src/models/persimmon.rs index 9657d419fb..0996decf55 100644 --- a/candle-transformers/src/models/persimmon.rs +++ b/candle-transformers/src/models/persimmon.rs @@ -1,6 +1,6 @@ -//! Persimmon Model, a transformer language model for efficient inference and general-purpose tasks +//! Persimmon Model //! -//! See Persimmon model details at: +//! A transformer language model for efficient inference and general-purpose tasks. See Persimmon model details at: //! - [Hugging Face](https://huggingface.co/adept/persimmon-8b-base) //! //! The model uses a standard transformer architecture with: diff --git a/candle-transformers/src/models/segment_anything/mod.rs b/candle-transformers/src/models/segment_anything/mod.rs index c54493d296..3e85fe3594 100644 --- a/candle-transformers/src/models/segment_anything/mod.rs +++ b/candle-transformers/src/models/segment_anything/mod.rs @@ -1,3 +1,11 @@ +//! Segment Anything Model (SAM) +//! +//! SAM is an architecture for image segmentation, capable of segmenting any object +//! in an image based on prompts like points or boxes. +//! +//! - [GH Link](https://github.com/facebookresearch/segment-anything) +//! - [Paper](https://arxiv.org/abs/2304.02643) +//! pub use crate::models::with_tracing::Linear; use candle::{Result, Tensor}; use candle_nn::{Module, VarBuilder};