Skip to content

Commit

Permalink
Switch to using llama-v2 by default. (#251)
Browse files Browse the repository at this point in the history
  • Loading branch information
LaurentMazare authored Jul 26, 2023
1 parent 89fd988 commit 84ad558
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions candle-examples/examples/llama/main.rs
Original file line number Diff line number Diff line change
Expand Up @@ -115,7 +115,7 @@ struct Args {
model_id: Option<String>,

#[arg(long)]
v2: bool,
v1: bool,

#[arg(long)]
use_flash_attn: bool,
Expand All @@ -139,10 +139,10 @@ fn main() -> Result<()> {
None => {
let api = Api::new()?;
let model_id = args.model_id.unwrap_or_else(|| {
if args.v2 {
"meta-llama/Llama-2-7b-hf".to_string()
} else {
if args.v1 {
"Narsil/amall-7b".to_string()
} else {
"meta-llama/Llama-2-7b-hf".to_string()
}
});
println!("loading the model weights from {model_id}");
Expand Down

0 comments on commit 84ad558

Please sign in to comment.