-
Notifications
You must be signed in to change notification settings - Fork 8
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
8622c74
commit f46c270
Showing
8 changed files
with
280 additions
and
43 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,97 @@ | ||
use candle_lora::LoraEmbeddingConfigBuilder; | ||
|
||
#[test] | ||
fn embed() -> candle_core::Result<()> { | ||
use std::{collections::HashMap, hash::Hash}; | ||
|
||
use candle_core::{DType, Device, Result, Tensor}; | ||
use candle_lora::{EmbeddingLayerLike, Lora, NewLayers, SelectedLayers}; | ||
use candle_nn::{init, Embedding, Module, VarMap}; | ||
|
||
#[derive(PartialEq, Eq, Hash)] | ||
enum ModelLayers { | ||
Embed, | ||
} | ||
|
||
#[derive(Debug)] | ||
struct Model { | ||
embed: Box<dyn EmbeddingLayerLike>, | ||
} | ||
|
||
impl Module for Model { | ||
fn forward(&self, input: &Tensor) -> Result<Tensor> { | ||
self.embed.forward(input) | ||
} | ||
} | ||
|
||
impl Model { | ||
fn insert_new(&mut self, new: NewLayers<ModelLayers>) { | ||
for (name, mut embed) in new.embed { | ||
match name { | ||
ModelLayers::Embed => { | ||
embed.merge().unwrap(); | ||
self.embed = Box::new(embed) | ||
} | ||
} | ||
} | ||
} | ||
} | ||
let device = Device::Cpu; | ||
let dtype = DType::F32; | ||
|
||
let in_size = 10; | ||
let hidden_size = 3; | ||
|
||
//Create the model | ||
let map = VarMap::new(); | ||
let embed_weight = map.get( | ||
(in_size, hidden_size), | ||
"embed.weight", | ||
init::ZERO, | ||
dtype, | ||
&device, | ||
)?; | ||
|
||
let mut model = Model { | ||
embed: Box::new(Embedding::new(embed_weight, hidden_size)), | ||
}; | ||
|
||
let dummy_image = Tensor::zeros((2, 4), DType::U32, &device)?; | ||
|
||
//Test the model | ||
let output = model.forward(&dummy_image).unwrap(); | ||
println!("Output: {output:?}"); | ||
|
||
//Select layers we want to convert | ||
let linear_layers = HashMap::new(); | ||
let conv1d_layers = HashMap::new(); | ||
let conv2d_layers = HashMap::new(); | ||
let mut embed_layers = HashMap::new(); | ||
embed_layers.insert(ModelLayers::Embed, &*model.embed); | ||
let selected = SelectedLayers { | ||
linear: linear_layers, | ||
linear_config: None, | ||
conv1d: conv1d_layers, | ||
conv1d_config: None, | ||
conv2d: conv2d_layers, | ||
conv2d_config: None, | ||
embed: embed_layers, | ||
embed_config: Some( | ||
LoraEmbeddingConfigBuilder::default(&device, dtype, in_size, hidden_size).build(), | ||
), | ||
}; | ||
|
||
//Create new LoRA layers from our layers | ||
let new_layers = Lora::convert_model(selected); | ||
|
||
//Custom methods to implement | ||
model.insert_new(new_layers); | ||
|
||
//Test the model | ||
let lora_output = model.forward(&dummy_image).unwrap(); | ||
println!("LoRA Output: {lora_output:?}"); | ||
|
||
assert_eq!(lora_output.shape(), output.shape()); | ||
|
||
Ok(()) | ||
} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.