Skip to content

Commit fef1a1c

Browse files
v0.4.3 (#152)
1 parent 84722f3 commit fef1a1c

File tree

8 files changed

+12
-16
lines changed

8 files changed

+12
-16
lines changed

Cargo.lock

Lines changed: 3 additions & 3 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

benchmark/Cargo.lock

Lines changed: 1 addition & 1 deletion
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

docs/openapi.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
"name": "Apache 2.0",
1212
"url": "https://www.apache.org/licenses/LICENSE-2.0"
1313
},
14-
"version": "0.4.2"
14+
"version": "0.4.3"
1515
},
1616
"paths": {
1717
"/generate": {

launcher/Cargo.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[package]
22
name = "text-generation-launcher"
3-
version = "0.4.2"
3+
version = "0.4.3"
44
edition = "2021"
55
authors = ["Olivier Dehaene"]
66
description = "Text Generation Launcher"

router/Cargo.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[package]
22
name = "text-generation-router"
3-
version = "0.4.2"
3+
version = "0.4.3"
44
edition = "2021"
55
authors = ["Olivier Dehaene"]
66
description = "Text Generation Webserver"

router/client/Cargo.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[package]
22
name = "text-generation-client"
3-
version = "0.4.2"
3+
version = "0.4.3"
44
edition = "2021"
55

66
[dependencies]

router/src/main.rs

Lines changed: 3 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ struct Args {
3737
max_waiting_tokens: usize,
3838
#[clap(default_value = "3000", long, short, env)]
3939
port: u16,
40-
#[clap(default_value = "/tmp/text-generation-server-0", long, env)]
40+
#[clap(default_value = "/tmp/text-generation-0", long, env)]
4141
master_shard_uds_path: String,
4242
#[clap(default_value = "bigscience/bloom", long, env)]
4343
tokenizer_name: String,
@@ -76,8 +76,6 @@ fn main() -> Result<(), std::io::Error> {
7676
panic!("validation_workers must be > 0");
7777
}
7878

79-
init_logging(otlp_endpoint, json_output);
80-
8179
// CORS allowed origins
8280
// map to go inside the option and then map to parse from String to HeaderValue
8381
// Finally, convert to AllowOrigin
@@ -91,28 +89,26 @@ fn main() -> Result<(), std::io::Error> {
9189

9290
// Tokenizer instance
9391
// This will only be used to validate payloads
94-
tracing::info!("Loading tokenizer");
9592
let local_path = Path::new(&tokenizer_name);
9693
let tokenizer =
9794
if local_path.exists() && local_path.is_dir() && local_path.join("tokenizer.json").exists()
9895
{
9996
// Load local tokenizer
100-
tracing::info!("Found local tokenizer");
10197
Tokenizer::from_file(local_path.join("tokenizer.json")).unwrap()
10298
} else {
10399
// Download and instantiate tokenizer
104100
// We need to download it outside of the Tokio runtime
105-
tracing::info!("Downloading tokenizer");
106101
Tokenizer::from_pretrained(tokenizer_name.clone(), None).unwrap()
107102
};
108-
tracing::info!("Tokenizer loaded");
109103

110104
// Launch Tokio runtime
111105
tokio::runtime::Builder::new_multi_thread()
112106
.enable_all()
113107
.build()
114108
.unwrap()
115109
.block_on(async {
110+
init_logging(otlp_endpoint, json_output);
111+
116112
// Get pipeline tag
117113
let model_info = reqwest::get(format!(
118114
"https://huggingface.co/api/models/{tokenizer_name}"

server/pyproject.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
[tool.poetry]
22
name = "text-generation-server"
3-
version = "0.4.2"
3+
version = "0.4.3"
44
description = "Text Generation Inference Python gRPC Server"
55
authors = ["Olivier Dehaene <olivier@huggingface.co>"]
66

0 commit comments

Comments
 (0)