Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FEAT: Vision 'img' option #108

Merged
merged 3 commits into from
Jun 19, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion commands.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@ commands:
prefix: '!'
provider: open-ai
type: quick-query
model: gpt-4-1106-preview
model: gpt-4o
traits:
- admin-only
- openai-moderated
Expand Down
3 changes: 3 additions & 0 deletions modules/builder.py
Original file line number Diff line number Diff line change
Expand Up @@ -60,6 +60,9 @@ class InvalidCommandException(Exception): ...
"allow-long",
"enable-hard-limit",
"hard-limit-length",
"allow-img",
"img-detail",
"img-screen-id"
}


Expand Down
37 changes: 33 additions & 4 deletions modules/conversation_history.py
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
import base64
from io import BytesIO
import mss
import mss.tools
from PIL import Image
from modules.lobby_manager import lobby_manager
from modules.typing import Message, MessageHistory
from modules.utils.prompts import PROMPTS, get_prompt_by_name
Expand All @@ -24,8 +29,8 @@ def _get_system_message(self) -> Message:

# Soft limiting the response
if (
self.settings.get("enable-soft-limit") is True
or self.settings.get("enable-soft-limit") is None
self.settings.get("enable-soft-limit") is True
or self.settings.get("enable-soft-limit") is None
):
enable_soft_limit = self.enable_soft_limit
else:
Expand Down Expand Up @@ -70,7 +75,7 @@ def add_assistant_message(self, message: Message) -> None:
self.message_history.append(message)

def add_user_message_from_prompt(
self, user_prompt: str, enable_soft_limit: bool = True
self, user_prompt: str, enable_soft_limit: bool = True
) -> None:
user_message = remove_args(user_prompt)
args = get_args(user_prompt)
Expand All @@ -87,10 +92,34 @@ def add_user_message_from_prompt(
if r"\stats" in args:
self.enable_stats = True

if r"\img" in args and self.settings.get("allow-img"):
sct = mss.mss()
monitor = sct.monitors[self.settings.get("img-screen-id", 1)]
scr = sct.grab(monitor)
img = Image.frombytes("RGB", scr.size, scr.bgra, "raw", "BGRX")
buffered = BytesIO()
img.save(buffered, format="JPEG")
base64_encoded_image = base64.b64encode(buffered.getvalue()).decode("utf-8")

content = [
{
"type": "text",
"text": f"{user_message}"
},
{
"type": "image_url",
"image_url": {
"url": f"data:image/jpeg;base64,{base64_encoded_image}",
"detail": self.settings.get("img-detail") or "low"
}
}]
else:
content = user_message

# Don't add a message if the user prompt is empty.
# Some LLM providers will complain about that, which effectively kills the chat.
if user_message != "":
self.message_history.append(Message(role="user", content=user_message))
self.message_history.append(Message(role="user", content=content))

def reset_turn(self):
self.enable_soft_limit = True
Expand Down
4 changes: 2 additions & 2 deletions modules/typing.py
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
from typing import Callable, List, Literal, NamedTuple, Optional, TypedDict
from typing import Callable, List, Literal, NamedTuple, Optional, TypedDict, Union, Dict

from pydantic import BaseModel


class Message(TypedDict):
role: Literal["assistant", "user", "system"]
content: str
content: Union[str, List[Dict[Literal["text", "image_url"], Union[str, Dict[str, str]]]]]


MessageHistory = List[Message]
Expand Down
3 changes: 2 additions & 1 deletion requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -18,4 +18,5 @@ pytest-repeat==0.9.3
groq==0.5.0
pyyaml==6.0.1
isort==5.13.2
black==24.4.2
black==24.4.2
mss==9.0.1
23 changes: 23 additions & 0 deletions schemas/commands.schema.json
Original file line number Diff line number Diff line change
Expand Up @@ -74,6 +74,19 @@
},
"hard-limit-length": {
"type": "integer"
},
"allow-img": {
"type": "boolean"
},
"img-screen-id": {
"type": "integer"
},
"img-detail": {
"type": "string",
"enum": [
"low",
"high"
]
}
},
"dependencies": {
Expand All @@ -86,6 +99,16 @@
"required": [
"hard-limit-length"
]
},
"img-screen-id": {
"required": [
"allow-img"
]
},
"img-detail": {
"required": [
"allow-img"
]
}
}
},
Expand Down
Loading