fix: should use spawn_blocking for workload without `yield`
parent
2d5b3e4ff5
commit
b69ad9a532
|
|
@ -690,7 +690,7 @@ dependencies = [
|
|||
|
||||
[[package]]
|
||||
name = "ctranslate2-bindings"
|
||||
version = "0.2.0"
|
||||
version = "0.2.2-rc.4"
|
||||
dependencies = [
|
||||
"async-stream",
|
||||
"async-trait",
|
||||
|
|
@ -3086,7 +3086,7 @@ dependencies = [
|
|||
|
||||
[[package]]
|
||||
name = "tabby"
|
||||
version = "0.2.1"
|
||||
version = "0.2.2-rc.4"
|
||||
dependencies = [
|
||||
"anyhow",
|
||||
"async-stream",
|
||||
|
|
|
|||
|
|
@ -13,7 +13,7 @@ members = [
|
|||
]
|
||||
|
||||
[workspace.package]
|
||||
version = "0.2.1"
|
||||
version = "0.2.2-rc.4"
|
||||
edition = "2021"
|
||||
authors = ["Meng Zhang"]
|
||||
homepage = "https://github.com/TabbyML/tabby"
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
[package]
|
||||
name = "ctranslate2-bindings"
|
||||
version = "0.2.0"
|
||||
version = "0.2.2-rc.4"
|
||||
edition = "2021"
|
||||
|
||||
[dependencies]
|
||||
|
|
|
|||
|
|
@ -129,29 +129,29 @@ impl TextGeneration for CTranslate2Engine {
|
|||
options: TextGenerationOptions,
|
||||
) -> BoxStream<String> {
|
||||
let encoding = self.tokenizer.encode(prompt, true).unwrap();
|
||||
let decoding = self.decoding_factory.create_incremental_decoding(
|
||||
self.tokenizer.clone(),
|
||||
truncate_tokens(encoding.get_ids(), options.max_input_length),
|
||||
options.stop_words,
|
||||
);
|
||||
|
||||
let cancel = CancellationToken::new();
|
||||
let engine = self.engine.clone();
|
||||
let (sender, mut receiver) = channel::<String>(8);
|
||||
let context = InferenceContext::new(sender, decoding, cancel.clone());
|
||||
tokio::task::spawn_blocking(move || {
|
||||
let context = Box::new(context);
|
||||
engine.inference(
|
||||
context,
|
||||
inference_callback,
|
||||
truncate_tokens(encoding.get_tokens(), options.max_input_length),
|
||||
options.max_decoding_length,
|
||||
options.sampling_temperature,
|
||||
);
|
||||
});
|
||||
|
||||
let s = stream! {
|
||||
let cancel = CancellationToken::new();
|
||||
let cancel_for_inference = cancel.clone();
|
||||
let _guard = cancel.drop_guard();
|
||||
|
||||
let decoding = self
|
||||
.decoding_factory
|
||||
.create_incremental_decoding(self.tokenizer.clone(), truncate_tokens(encoding.get_ids(), options.max_input_length), options.stop_words);
|
||||
|
||||
let (sender, mut receiver) = channel::<String>(8);
|
||||
let context = InferenceContext::new(sender, decoding, cancel_for_inference);
|
||||
tokio::task::spawn(async move {
|
||||
let context = Box::new(context);
|
||||
engine.inference(
|
||||
context,
|
||||
inference_callback,
|
||||
truncate_tokens(encoding.get_tokens(), options.max_input_length),
|
||||
options.max_decoding_length,
|
||||
options.sampling_temperature,
|
||||
);
|
||||
});
|
||||
|
||||
while let Some(text) = receiver.recv().await {
|
||||
yield text;
|
||||
}
|
||||
|
|
|
|||
|
|
@ -1,6 +1,6 @@
|
|||
[package]
|
||||
name = "tabby"
|
||||
version = "0.2.1"
|
||||
version = "0.2.2-rc.4"
|
||||
edition = "2021"
|
||||
|
||||
[dependencies]
|
||||
|
|
|
|||
Loading…
Reference in New Issue