AppFlowy/frontend/rust-lib/flowy-chat/src/event_map.rs
Nathan.fooo c6ad57f11d
chore: show plugin state (#5740)
* chore: show plugin state

* chore: flutter analyzer

* chore: update
2024-07-15 22:45:53 +08:00

99 lines
3.1 KiB
Rust

use std::sync::{Arc, Weak};
use strum_macros::Display;
use crate::tools::AITools;
use flowy_derive::{Flowy_Event, ProtoBuf_Enum};
use lib_dispatch::prelude::*;
use crate::chat_manager::ChatManager;
use crate::event_handler::*;
pub fn init(chat_manager: Weak<ChatManager>) -> AFPlugin {
let user_service = Arc::downgrade(&chat_manager.upgrade().unwrap().user_service);
let cloud_service = Arc::downgrade(&chat_manager.upgrade().unwrap().chat_service_wm);
let ai_tools = Arc::new(AITools::new(cloud_service, user_service));
AFPlugin::new()
.name("Flowy-Chat")
.state(chat_manager)
.state(ai_tools)
.event(ChatEvent::StreamMessage, stream_chat_message_handler)
.event(ChatEvent::LoadPrevMessage, load_prev_message_handler)
.event(ChatEvent::LoadNextMessage, load_next_message_handler)
.event(ChatEvent::GetRelatedQuestion, get_related_question_handler)
.event(ChatEvent::GetAnswerForQuestion, get_answer_handler)
.event(ChatEvent::StopStream, stop_stream_handler)
.event(
ChatEvent::RefreshLocalAIModelInfo,
refresh_local_ai_info_handler,
)
.event(ChatEvent::UpdateLocalLLM, update_local_llm_model_handler)
.event(ChatEvent::GetLocalLLMState, get_local_llm_state_handler)
.event(ChatEvent::CompleteText, start_complete_text_handler)
.event(ChatEvent::StopCompleteText, stop_complete_text_handler)
.event(ChatEvent::ChatWithFile, chat_file_handler)
.event(
ChatEvent::DownloadLLMResource,
download_llm_resource_handler,
)
.event(
ChatEvent::CancelDownloadLLMResource,
cancel_download_llm_resource_handler,
)
.event(ChatEvent::GetPluginState, get_plugin_state_handler)
.event(ChatEvent::RestartLocalAI, restart_local_ai_handler)
}
#[derive(Clone, Copy, PartialEq, Eq, Debug, Display, Hash, ProtoBuf_Enum, Flowy_Event)]
#[event_err = "FlowyError"]
pub enum ChatEvent {
/// Create a new workspace
#[event(input = "LoadPrevChatMessagePB", output = "ChatMessageListPB")]
LoadPrevMessage = 0,
#[event(input = "LoadNextChatMessagePB", output = "ChatMessageListPB")]
LoadNextMessage = 1,
#[event(input = "StreamChatPayloadPB", output = "ChatMessagePB")]
StreamMessage = 2,
#[event(input = "StopStreamPB")]
StopStream = 3,
#[event(input = "ChatMessageIdPB", output = "RepeatedRelatedQuestionPB")]
GetRelatedQuestion = 4,
#[event(input = "ChatMessageIdPB", output = "ChatMessagePB")]
GetAnswerForQuestion = 5,
#[event(input = "LLMModelPB", output = "LocalModelResourcePB")]
UpdateLocalLLM = 6,
#[event(output = "LocalModelResourcePB")]
GetLocalLLMState = 7,
#[event(output = "LLMModelInfoPB")]
RefreshLocalAIModelInfo = 8,
#[event(input = "CompleteTextPB", output = "CompleteTextTaskPB")]
CompleteText = 9,
#[event(input = "CompleteTextTaskPB")]
StopCompleteText = 10,
#[event(input = "ChatFilePB")]
ChatWithFile = 11,
#[event(input = "DownloadLLMPB", output = "DownloadTaskPB")]
DownloadLLMResource = 12,
#[event()]
CancelDownloadLLMResource = 13,
#[event(output = "PluginStatePB")]
GetPluginState = 14,
#[event()]
RestartLocalAI = 15,
}