feat: add hugging chat mask
This commit is contained in:
parent
5d073ab5c4
commit
bc2fab30a8
19
src/main.py
19
src/main.py
|
@ -1126,14 +1126,19 @@ Close All Without Dialog: {self.close_all_without_dialog}
|
||||||
|
|
||||||
self.win.banner.set_revealed(False)
|
self.win.banner.set_revealed(False)
|
||||||
|
|
||||||
for an in self.annoucements.values():
|
for key, an in self.annoucements.items():
|
||||||
if an["provider"] == self.provider:
|
if an["provider"] == self.provider:
|
||||||
if an["status"] == "open":
|
if an["status"] == "open":
|
||||||
self.win.banner.set_title(an["message"])
|
match an["action"]:
|
||||||
self.win.banner.props.button_label = "Open settings"
|
case "error": # show an error banner with a button to open settings
|
||||||
self.win.banner.connect("button-clicked", self.on_preferences_action)
|
self.win.banner.set_title(an["message"])
|
||||||
self.win.banner.set_revealed(True)
|
self.win.banner.props.button_label = "Open settings"
|
||||||
return
|
self.win.banner.connect("button-clicked", self.on_preferences_action)
|
||||||
|
self.win.banner.set_revealed(True)
|
||||||
|
case _:
|
||||||
|
self.win.banner.set_title(an["message"])
|
||||||
|
self.win.banner.set_revealed(True)
|
||||||
|
del self.annoucements[key]
|
||||||
break
|
break
|
||||||
|
|
||||||
self.prompt = self.win.prompt_text_view.get_buffer().props.text.strip()
|
self.prompt = self.win.prompt_text_view.get_buffer().props.text.strip()
|
||||||
|
@ -1151,6 +1156,8 @@ Close All Without Dialog: {self.close_all_without_dialog}
|
||||||
response = self.providers[self.provider].ask(self.prompt)
|
response = self.providers[self.provider].ask(self.prompt)
|
||||||
except GLib.Error as e:
|
except GLib.Error as e:
|
||||||
response = e.message
|
response = e.message
|
||||||
|
except KeyError:
|
||||||
|
del self.providers[self.provider]
|
||||||
GLib.idle_add(cleanup, response)
|
GLib.idle_add(cleanup, response)
|
||||||
|
|
||||||
def cleanup(response):
|
def cleanup(response):
|
||||||
|
|
|
@ -72,6 +72,7 @@ class Preferences(Adw.PreferencesWindow):
|
||||||
"""Callback for the allow_remote_fetching_switch toggled event."""
|
"""Callback for the allow_remote_fetching_switch toggled event."""
|
||||||
if widget.get_active():
|
if widget.get_active():
|
||||||
self.settings.set_boolean("allow-remote-fetching", True)
|
self.settings.set_boolean("allow-remote-fetching", True)
|
||||||
|
self.app.load_annoucements()
|
||||||
self.app.allow_remote_fetching = True
|
self.app.allow_remote_fetching = True
|
||||||
else:
|
else:
|
||||||
self.settings.set_boolean("allow-remote-fetching", False)
|
self.settings.set_boolean("allow-remote-fetching", False)
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
#from .huggingchat import HuggingChatProvider
|
# from .huggingchat import HuggingChatProvider
|
||||||
from .baichat import BAIChatProvider
|
from .baichat import BAIChatProvider
|
||||||
from .openaigpt35turbo import OpenAIGPT35TurboProvider
|
from .openaigpt35turbo import OpenAIGPT35TurboProvider
|
||||||
from .openaigpt4 import OpenAIGPT4Provider
|
from .openaigpt4 import OpenAIGPT4Provider
|
||||||
|
@ -7,13 +7,13 @@ from .openaitextdavinci003 import OpenAITextDavinci003
|
||||||
from .alpacalora import AlpacaLoRAProvider
|
from .alpacalora import AlpacaLoRAProvider
|
||||||
from .hfgoogleflant5xxl import HuggingFaceGoogleFlanT5XXLProvider
|
from .hfgoogleflant5xxl import HuggingFaceGoogleFlanT5XXLProvider
|
||||||
from .hfgoogleflanu12 import HuggingFaceGoogleFlanU12Provider
|
from .hfgoogleflanu12 import HuggingFaceGoogleFlanU12Provider
|
||||||
from .hfopenassistantsft1pythia12b import HuggingFaceOpenAssistantSFT1PythiaProvider
|
from .hfopenassistantsft1pythia12b import HuggingFaceOpenAssistantSFT1PythiaProvider, HuggingChatMask
|
||||||
from .hfgpt2 import HuggingFaceGPT2Provider
|
from .hfgpt2 import HuggingFaceGPT2Provider
|
||||||
from .hfdialogpt import HuggingFaceDialoGPTLargeProvider
|
from .hfdialogpt import HuggingFaceDialoGPTLargeProvider
|
||||||
from .bard import BardProvider
|
# from .bard import BardProvider
|
||||||
from .hfgpt2large import HuggingFaceGPT2LargeProvider
|
from .hfgpt2large import HuggingFaceGPT2LargeProvider
|
||||||
from .hfgpt2xl import HuggingFaceGPT2XLProvider
|
from .hfgpt2xl import HuggingFaceGPT2XLProvider
|
||||||
#from .starcoder import StarCoderProvider
|
# from .starcoder import StarCoderProvider
|
||||||
|
|
||||||
PROVIDERS = {
|
PROVIDERS = {
|
||||||
"alpacalora": AlpacaLoRAProvider,
|
"alpacalora": AlpacaLoRAProvider,
|
||||||
|
@ -27,7 +27,7 @@ PROVIDERS = {
|
||||||
"hfgpt2large": HuggingFaceGPT2LargeProvider,
|
"hfgpt2large": HuggingFaceGPT2LargeProvider,
|
||||||
"hfgpt2xl": HuggingFaceGPT2XLProvider,
|
"hfgpt2xl": HuggingFaceGPT2XLProvider,
|
||||||
"hfopenassistantsft1pythia12b": HuggingFaceOpenAssistantSFT1PythiaProvider,
|
"hfopenassistantsft1pythia12b": HuggingFaceOpenAssistantSFT1PythiaProvider,
|
||||||
"huggingchat": HuggingFaceOpenAssistantSFT1PythiaProvider, # hugging chat is replaced by open assistant
|
"huggingchat": HuggingChatMask, # hugging chat is replaced by open assistant
|
||||||
"openaigpt35turbo": OpenAIGPT35TurboProvider,
|
"openaigpt35turbo": OpenAIGPT35TurboProvider,
|
||||||
"openaigpt4": OpenAIGPT4Provider,
|
"openaigpt4": OpenAIGPT4Provider,
|
||||||
"openaitextdavinci003": OpenAITextDavinci003,
|
"openaitextdavinci003": OpenAITextDavinci003,
|
||||||
|
|
|
@ -46,4 +46,9 @@ class HuggingFaceOpenAssistantSFT1PythiaProvider(BaseHFProvider):
|
||||||
self.hide_banner()
|
self.hide_banner()
|
||||||
print(response)
|
print(response)
|
||||||
GLib.idle_add(self.update_response, response)
|
GLib.idle_add(self.update_response, response)
|
||||||
return response
|
return response
|
||||||
|
|
||||||
|
class HuggingChatMask(HuggingFaceOpenAssistantSFT1PythiaProvider):
|
||||||
|
url = "https://bavarder.codeberg.page/help/huggingchat"
|
||||||
|
name = "Open-Assistant SFT-1 12B Model (HuggingChat)"
|
||||||
|
slug = "huggingchat"
|
Loading…
Reference in New Issue
Block a user