-
Notifications
You must be signed in to change notification settings - Fork 3
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
9 changed files
with
497 additions
and
484 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,19 +1,42 @@ | ||
[build-system] | ||
requires = ["setuptools>=57.0.0", "wheel"] | ||
build-backend = "setuptools.build_meta" | ||
|
||
[project] | ||
name = "jetson-examples" | ||
version = "0.0.3" | ||
authors = [{ name = "luozhixin", email = "[email protected]" }] | ||
description = "run AI on Jetson, all you need is reComputer" | ||
description = "Running Gen AI models and applications on NVIDIA Jetson devices with one-line command" | ||
readme = "README.md" | ||
requires-python = ">=3.8" | ||
classifiers = [ | ||
"Programming Language :: Python :: 3", | ||
"License :: OSI Approved :: MIT License", | ||
"Operating System :: OS Independent", | ||
] | ||
keywords = [ | ||
"llama", | ||
"llava", | ||
"gpt", | ||
"llm", | ||
"nvidia", | ||
"jetson", | ||
"multimodal", | ||
"jetson orin", | ||
] | ||
|
||
[project.scripts] | ||
reComputer = "reComputer.main:run_script" | ||
|
||
[project.urls] | ||
Homepage = "https://github.com/Seeed-Projects/jetson-examples" | ||
Issues = "https://github.com/Seeed-Projects/jetson-examples/issues" | ||
|
||
# Tools settings ------------------------------------------------------------------------------------------------------- | ||
|
||
[tool.setuptools.packages.find] | ||
namespaces = true | ||
where = ["."] | ||
|
||
[tool.setuptools.package-data] | ||
"reComputer.scripts" = ["*.sh"] |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1 @@ | ||
__version__ = "0.0.4" |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,5 @@ | ||
script_dir=$(dirname "$0") | ||
docker --version && \ | ||
python3 -V && \ | ||
python -V && \ | ||
echo "now we can use more shell in $script_dir" |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,278 @@ | ||
#!/bin/bash | ||
|
||
SUPPORT_L4T_LIST="35.3.1" | ||
BASE_PATH=/home/$USER/reComputer | ||
JETSON_REPO_PATH="$BASE_PATH/jetson-containers" | ||
|
||
get_l4t_version() { | ||
ARCH=$(uname -i) | ||
echo "ARCH: $ARCH" | ||
|
||
if [ $ARCH = "aarch64" ]; then | ||
L4T_VERSION_STRING=$(head -n 1 /etc/nv_tegra_release) | ||
|
||
if [ -z "$L4T_VERSION_STRING" ]; then | ||
echo "reading L4T version from \"dpkg-query --show nvidia-l4t-core\"" | ||
L4T_VERSION_STRING=$(dpkg-query --showformat='${Version}' --show nvidia-l4t-core) | ||
L4T_VERSION_ARRAY=(${L4T_VERSION_STRING//./ }) | ||
L4T_RELEASE=${L4T_VERSION_ARRAY[0]} | ||
L4T_REVISION=${L4T_VERSION_ARRAY[1]} | ||
else | ||
echo "reading L4T version from /etc/nv_tegra_release" | ||
L4T_RELEASE=$(echo $L4T_VERSION_STRING | cut -f 2 -d ' ' | grep -Po '(?<=R)[^;]+') | ||
L4T_REVISION=$(echo $L4T_VERSION_STRING | cut -f 2 -d ',' | grep -Po '(?<=REVISION: )[^;]+') | ||
fi | ||
|
||
L4T_REVISION_MAJOR=${L4T_REVISION:0:1} | ||
L4T_REVISION_MINOR=${L4T_REVISION:2:1} | ||
L4T_VERSION="$L4T_RELEASE.$L4T_REVISION" | ||
|
||
echo "L4T_VERSION: $L4T_VERSION" | ||
|
||
elif [ $ARCH != "x86_64" ]; then | ||
echo "unsupported architecture: $ARCH" # show in red color | ||
exit 1 | ||
fi | ||
} | ||
|
||
# 1. Check L4T version | ||
get_l4t_version | ||
CHECK_L4T_VERSION=0 | ||
for item in $SUPPORT_L4T_LIST; do | ||
if [ "$item" = "$L4T_VERSION" ]; then | ||
CHECK_L4T_VERSION=1 | ||
break | ||
fi | ||
done | ||
|
||
if [ $CHECK_L4T_VERSION -eq 1 ]; then | ||
echo "pass the version check" | ||
else | ||
echo "currently supported versions of jetpack are $SUPPORT_L4T_LIST" # show in red color | ||
exit 1 | ||
fi | ||
|
||
# 2. Check Google Chrome | ||
if dpkg -s chromium-browser &>/dev/null; then | ||
echo "Chrome is installed." | ||
else | ||
echo "install Google Chrome ..." # show in red color | ||
sudo apt install chromium-browser | ||
echo "Google Chrome installed successfully" # show in red color | ||
fi | ||
|
||
# 3. Generate Google browser key | ||
FILE_NAME="key.pem" | ||
FILE_PATH="$JETSON_REPO_PATH/data" | ||
if [ -f "$FILE_PATH/$FILE_NAME" ]; then | ||
echo "key file '$FILE_PATH/$FILE_NAME' exists." | ||
else | ||
cd $FILE_PATH | ||
openssl req -x509 -newkey rsa:4096 -keyout key.pem -out cert.pem -sha256 -days 365 -nodes -subj '/CN=localhost' | ||
cd .. | ||
fi | ||
|
||
# 4. edit source code | ||
cat >"$JETSON_REPO_PATH/packages/llm/local_llm/agents/video_query.py" <<'EOF' | ||
#!/usr/bin/env python3 | ||
import time | ||
import logging | ||
import threading | ||
from local_llm import Agent | ||
from local_llm.plugins import ( | ||
VideoSource, | ||
VideoOutput, | ||
ChatQuery, | ||
PrintStream, | ||
ProcessProxy, | ||
) | ||
from local_llm.utils import ArgParser, print_table | ||
from termcolor import cprint | ||
from jetson_utils import cudaFont, cudaMemcpy, cudaToNumpy, cudaDeviceSynchronize | ||
from flask import Flask, request | ||
class VideoQuery(Agent): | ||
""" | ||
Perpetual always-on closed-loop visual agent that applies prompts to a video stream. | ||
""" | ||
def __init__(self, model="liuhaotian/llava-v1.5-7b", **kwargs): | ||
super().__init__() | ||
self.lock = threading.Lock() | ||
# load model in another process for smooth streaming | ||
# self.llm = ProcessProxy((lambda **kwargs: ChatQuery(model, drop_inputs=True, **kwargs)), **kwargs) | ||
self.llm = ChatQuery(model, drop_inputs=True, **kwargs) | ||
self.llm.add(PrintStream(color="green", relay=True).add(self.on_text)) | ||
self.llm.start() | ||
# test / warm-up query | ||
self.warmup = True | ||
self.text = "" | ||
self.eos = False | ||
self.llm("What is 2+2?") | ||
while self.warmup: | ||
time.sleep(0.25) | ||
# create video streams | ||
self.video_source = VideoSource(**kwargs) | ||
self.video_output = VideoOutput(**kwargs) | ||
self.video_source.add(self.on_video, threaded=False) | ||
self.video_output.start() | ||
self.font = cudaFont() | ||
# setup prompts | ||
self.prompt = "Describe the image concisely and briefly." | ||
# entry node | ||
self.pipeline = [self.video_source] | ||
def on_video(self, image): | ||
np_image = cudaToNumpy(image) | ||
cudaDeviceSynchronize() | ||
self.llm( | ||
[ | ||
"reset", | ||
np_image, | ||
self.prompt, | ||
] | ||
) | ||
text = self.text.replace("\n", "").replace("</s>", "").strip() | ||
if text: | ||
worlds = text.split() | ||
line_counter = len(worlds) // 10 | ||
if len(worlds) % 10 != 0: | ||
line_counter += 1 | ||
for l in range(line_counter): | ||
line_text = " ".join(worlds[l * 10 : (l + 1) * 10]) | ||
self.font.OverlayText( | ||
image, | ||
text=line_text, | ||
x=5, | ||
y=int(79 + l * 37), | ||
color=self.font.White, | ||
background=self.font.Gray40, | ||
) | ||
self.font.OverlayText( | ||
image, | ||
text="Prompt: " + self.prompt, | ||
x=5, | ||
y=42, | ||
color=(120, 215, 21), | ||
background=self.font.Gray40, | ||
) | ||
self.video_output(image) | ||
def on_text(self, text): | ||
if self.eos: | ||
self.text = text # new query response | ||
self.eos = False | ||
elif not self.warmup: # don't view warmup response | ||
self.text = self.text + text | ||
if text.endswith("</s>") or text.endswith("###") or text.endswith("<|im_end|>"): | ||
self.print_stats() | ||
self.warmup = False | ||
self.eos = True | ||
def update_switch(self, on_off): | ||
self.video_source.switch(on_off) | ||
def update_prompts(self, new_prompt): | ||
with self.lock: | ||
if new_prompt: | ||
self.prompt = new_prompt | ||
def print_stats(self): | ||
# print_table(self.llm.model.stats) | ||
curr_time = time.perf_counter() | ||
if not hasattr(self, "start_time"): | ||
self.start_time = curr_time | ||
else: | ||
frame_time = curr_time - self.start_time | ||
self.start_time = curr_time | ||
logging.info( | ||
f"refresh rate: {1.0 / frame_time:.2f} FPS ({frame_time*1000:.1f} ms)" | ||
) | ||
if __name__ == "__main__": | ||
parser = ArgParser(extras=ArgParser.Defaults + ["video_input", "video_output"]) | ||
args = parser.parse_args() | ||
# 独立线程运行 | ||
agent = VideoQuery(**vars(args)) | ||
def run_video_query(): | ||
agent.run() | ||
video_query_thread = threading.Thread(target=run_video_query) | ||
video_query_thread.start() | ||
# 启动web服务 | ||
app = Flask(__name__) | ||
@app.route("/update_prompt", methods=["POST"]) | ||
def update_prompts(): | ||
prompt = request.json.get("prompt") | ||
if prompt: | ||
agent.update_prompts(prompt) | ||
return "Prompts updated successfully." | ||
else: | ||
return "Invalid prompts data." | ||
@app.route("/update_switch", methods=["POST"]) | ||
def update_switch(): | ||
infer_or_not = True if request.json.get("switch") == "on" else False | ||
agent.update_switch(infer_or_not) | ||
return "stop" if not infer_or_not else "start" | ||
@app.route("/update_params", methods=["POST"]) | ||
def update_params(): | ||
try: | ||
agent.llm.max_new_tokens = request.json.get("max_new_tokens") or 128 | ||
agent.llm.min_new_tokens = request.json.get("min_new_tokens") or -1 | ||
agent.llm.do_sample = request.json.get("do_sample") or False | ||
agent.llm.repetition_penalty = request.json.get("repetition_penalty") or 1.0 | ||
agent.llm.temperature = request.json.get("temperature") or 0.7 | ||
agent.llm.top_p = request.json.get("top_p") or 0.95 | ||
if request.json.get("system_prompt"): | ||
agent.llm.chat_history.template["system_prompt"] = request.json.get( | ||
"system_prompt" | ||
) | ||
return "params updated." | ||
except Exception as e: | ||
print(e) | ||
return "update failure" | ||
app.run(host="0.0.0.0", port=5555) | ||
EOF | ||
|
||
sed -i 's/from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection, SiglipImageProcessor, SiglipVisionModel/from transformers import CLIPImageProcessor, CLIPVisionModelWithProjection # , SiglipImageProcessor, SiglipVisionModel/' "$JETSON_REPO_PATH/packages/llm/local_llm/vision/clip_hf.py" | ||
sed -i "s/'siglip': dict(preprocessor=SiglipImageProcessor, model=SiglipVisionModel),/# 'siglip': dict(preprocessor=SiglipImageProcessor, model=SiglipVisionModel),/" "$JETSON_REPO_PATH/packages/llm/local_llm/vision/clip_hf.py" | ||
|
||
sed -i 's/from .audio import */# from .audio import */' "$JETSON_REPO_PATH/packages/llm/local_llm/plugins/__init__.py" | ||
sed -i 's/from .nanodb import NanoDB/# from .nanodb import NanoDB/' "$JETSON_REPO_PATH/packages/llm/local_llm/plugins/__init__.py" | ||
|
||
sed -i 's/import onnxruntime as ort/# import onnxruntime as ort/' "$JETSON_REPO_PATH/packages/llm/local_llm/utils/model.py" | ||
|
||
echo "The script has been modified." | ||
|
||
# gnome-terminal -- /bin/bash -c chromium-browser --disable-features=WebRtcHideLocalIpsWithMdns https://localhost:8554/"; exec /bin/bash" | ||
|
||
cd $JETSON_REPO_PATH | ||
sudo docker run --runtime nvidia -it --rm --network host --volume /tmp/argus_socket:/tmp/argus_socket --volume /etc/enctune.conf:/etc/enctune.conf --volume /etc/nv_tegra_release:/etc/nv_tegra_release --volume /proc/device-tree/model:/tmp/nv_jetson_model --volume /var/run/dbus:/var/run/dbus --volume /var/run/avahi-daemon/socket:/var/run/avahi-daemon/socket --volume /var/run/docker.sock:/var/run/docker.sock --volume $JETSON_REPO_PATH/data:/data --device /dev/snd --device /dev/bus/usb -e DISPLAY=:0 -v /tmp/.X11-unix/:/tmp/.X11-unix -v /tmp/.docker.xauth:/tmp/.docker.xauth -e XAUTHORITY=/tmp/.docker.xauth --device /dev/video0 --device /dev/video1 -v $JETSON_REPO_PATH/packages/llm/local_llm:/opt/local_llm/local_llm -e SSL_KEY=/data/key.pem -e SSL_CERT=/data/cert.pem dustynv/local_llm:r35.3.1 python3 -m local_llm.agents.video_query --api=mlc --verbose --model liuhaotian/llava-v1.5-7b --max-new-tokens 32 --video-input /dev/video0 --video-output webrtc:https://@:8554/output |
Oops, something went wrong.