Skip to content

Commit

Permalink
🏷️ Update based on open-sourced types (#185)
Browse files Browse the repository at this point in the history
* 🏷️ Update based on open-sourced types

Signed-off-by: Evaline Ju <[email protected]>

* 💡 Add caikit reference as comments

Signed-off-by: Evaline Ju <[email protected]>

---------

Signed-off-by: Evaline Ju <[email protected]>
  • Loading branch information
evaline-ju authored Sep 6, 2024
1 parent 4d136ad commit f970d56
Show file tree
Hide file tree
Showing 5 changed files with 19 additions and 23 deletions.
12 changes: 11 additions & 1 deletion protos/caikit_data_model_nlp.proto
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@

// Ref. https://github.com/caikit/caikit/blob/main/caikit/interfaces/nlp/data_model/text.py
/*------------------------------------------------------------------------------
* AUTO GENERATED
*----------------------------------------------------------------------------*/
Expand Down Expand Up @@ -27,6 +27,16 @@

/*-- MESSAGES ----------------------------------------------------------------*/

message ChunkerTokenizationStreamResult {

/*-- fields --*/
repeated caikit_data_model.nlp.Token results = 1;
int64 token_count = 4;
int64 processed_index = 2;
int64 start_index = 3;
int64 input_start_index = 20;
int64 input_end_index = 21;
}

message GeneratedTextResult {

Expand Down
17 changes: 3 additions & 14 deletions protos/caikit_runtime_Chunkers.proto
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@

// Ref. https://github.com/caikit/caikit/blob/main/caikit/interfaces/nlp/tasks.py
/*------------------------------------------------------------------------------
* AUTO GENERATED
*----------------------------------------------------------------------------*/
Expand All @@ -23,21 +23,10 @@ message ChunkerTokenizationTaskRequest {
string text = 1;
}

message ChunkerTokenizationStreamResult {

/*-- fields --*/
repeated caikit_data_model.nlp.Token results = 1;
int64 token_count = 4;
int64 processed_index = 2;
int64 start_index = 3;
int64 input_start_index = 5;
int64 input_end_index = 6;
}


/*-- SERVICES ----------------------------------------------------------------*/

service ChunkersService {
rpc BidiStreamingChunkerTokenizationTaskPredict(stream BidiStreamingChunkerTokenizationTaskRequest) returns (stream ChunkerTokenizationStreamResult);
rpc ChunkerTokenizationTaskPredict(ChunkerTokenizationTaskRequest) returns (caikit_data_model.nlp.TokenizationResults);
rpc BidiStreamingChunkerTokenizationTaskPredict(stream caikit.runtime.Chunkers.BidiStreamingChunkerTokenizationTaskRequest) returns (stream caikit_data_model.nlp.ChunkerTokenizationStreamResult);
rpc ChunkerTokenizationTaskPredict(caikit.runtime.Chunkers.ChunkerTokenizationTaskRequest) returns (caikit_data_model.nlp.TokenizationResults);
}
5 changes: 2 additions & 3 deletions src/clients/chunker.rs
Original file line number Diff line number Diff line change
Expand Up @@ -30,10 +30,9 @@ use crate::{
pb::{
caikit::runtime::chunkers::{
chunkers_service_client::ChunkersServiceClient,
BidiStreamingChunkerTokenizationTaskRequest, ChunkerTokenizationStreamResult,
ChunkerTokenizationTaskRequest,
BidiStreamingChunkerTokenizationTaskRequest, ChunkerTokenizationTaskRequest,
},
caikit_data_model::nlp::{Token, TokenizationResults},
caikit_data_model::nlp::{ChunkerTokenizationStreamResult, Token, TokenizationResults},
},
};

Expand Down
3 changes: 2 additions & 1 deletion src/orchestrator/streaming.rs
Original file line number Diff line number Diff line change
Expand Up @@ -37,7 +37,8 @@ use crate::{
unary::{input_detection_task, tokenize},
UNSUITABLE_INPUT_MESSAGE,
},
pb::caikit::runtime::chunkers::{self, ChunkerTokenizationStreamResult},
pb::caikit::runtime::chunkers,
pb::caikit_data_model::nlp::ChunkerTokenizationStreamResult,
};

pub type Chunk = ChunkerTokenizationStreamResult;
Expand Down
5 changes: 1 addition & 4 deletions src/orchestrator/streaming/aggregator.rs
Original file line number Diff line number Diff line change
Expand Up @@ -446,10 +446,7 @@ mod tests {
use super::*;
use crate::{
models::TokenClassificationResult,
pb::{
caikit::runtime::chunkers::ChunkerTokenizationStreamResult,
caikit_data_model::nlp::Token,
},
pb::caikit_data_model::nlp::{ChunkerTokenizationStreamResult, Token},
};

fn get_detection_obj(
Expand Down

0 comments on commit f970d56

Please sign in to comment.