Skip to content

relston/mark

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

90 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Build Status License

Mark

Mark lets you seamlessly use markdown, images and links to interact with GPT4o

"Markdown is still the best format for feeding into language models"
- Simon Willison (co-creator of Django)

Why Markdown + GPT? For a full rundown of the benefits and use cases, check out Introducing 'Mark', a Markdown CLI tool for GPT4o

Key Features

  • Interact with LLMs using Markdown
  • Visual recognition of markdown image references via GPT4o
  • Local and remote links are scraped for context
  • GPT responses appended directly into Markdown files
  • stdin and stdout support for piping
  • Support for Image Generation

Example

Describe ![this image](https://www.animal.photos/mamm1/cat-tig1_files/bengal12.jpg) to me in one short sentence.

# GPT Response (model: gpt-4o-2024-05-13, system: default)
A Bengal tiger is walking on a sandy path with greenery in the background.

# User Response
List each section of content on [this page](https://en.wikipedia.org/wiki/Bengal_tiger)

# GPT Response (model: gpt-4o-2024-05-13, system: default)
## Sections of Content on the Bengal Tiger Wikipedia Page

1. **Taxonomy**
   - Genetic ancestry

2. **Characteristics**
   - Body weight and size

3. **Distribution and habitat**
   - India
   - Bangladesh
    ...

This allows you to carry on a conversation directly in the markdown file - or simply integrate the response back into the document you're working on. This is a powerful, flexible and natural way to interact with LLMs that unlocks countless use cases. mark's focus on explicitly using Markdown as both a prompt and response medium for LLMs presents a unique approach. By blending interactive, version-controlled Markdown documents with LLM responses in a CLI tool offers an innovative workflow for those looking to integrate LLM interactions into their development and documentation processes.

Install the Mark CLI

pipx install git+https://github.com/relston/mark.git

Updating the CLI:

pipx upgrade mark

Usage

By default, mark will read a markdown file, extract any context references, and send them to the LLM. The responses are then appended to the markdown file.

mark path/to/markdown.md

Requires an OpenAI API key in the OPENAI_API_KEY environment variable

Also supports stdin with stdout for piping GPT responses into other tools

cat path/to/markdown.md | mark 
# LLM response....

Use a specific LLM model

You can specify a different LLM model to use with the --model (or -m) flag

mark path/to/markdown.md --model gpt-4o-2024-05-13

Custom system prompts

The system prompts folder is located at ~/.mark/system_prompts and it includes a default.md prompt. You can add any additional system prompts you'd like to use in this folder and use them with the --system (or -s) flag.

# ~/.mark/system_prompts/custom.md
mark path/to/markdown.md --system custom

Override the OpenAI API endpoint

If you want to use a different LLM API endpoint that is fully compatible with the OpenAI API, set the OPENAI_API_BASE_URL environment variable to that endpoint value. This should enable you to use OpenAI proxy services like credal.ai, or other LLMs that are compatible with the OpenAI SDK.

Image Generation

To generate an image based on the input just add the --generate-image (or -i) flag to the command

mark path/to/markdown.md --generate-image

This will generate an image using the 'dall-e-3' model and append it to the markdown file.

Development

Local Setup

poetry install

Requires poetry

Run the CLI Tool locally

poetry run mark path/to/markdown.md

Run the tests

poetry run python -m pytest

Auto-fix lint errors

poetry run autopep8 --in-place --aggressive --aggressive --recursive .