formulaic-ai


Nameformulaic-ai JSON
Version 0.1.0.4 PyPI version JSON
download
home_pageNone
SummaryThe official Formulaic python library for working with generative AI prompts.
upload_time2024-03-25 22:02:22
maintainerNone
docs_urlNone
authorNone
requires_python>=3.8
licenseNone
keywords llm api generative ai language models openai prompt engineering prompt scripts
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # Formulaic Python Library 
_**NOTE: This is a project in active development. It is not stable and changes frequently. It is not ready for production use.**_

The Formulaic library makes it easy to use Formulas inside your generative AI applications. Formulas are open-licensed JSON scripts that contain AI prompts, template variables, and model configuration.  You can browse the library of existing Formulas for many popular language models at [Formulaic.app](https://formulaic.app). 

## Installation
This repo is currently private inside Mozilla until we release it as open source. 

### (Recommended) Create a virtual environment
Create a new directory on your local machine and open a terminal in that directory. Run this command:

```
python -m venv venv
```

Note, on your machine you may have to specify python3 if "python" is not setup as a shortcut to your current version:

```
python3 -m venv venv
```

This command creates a 'venv' virtual environment under `./venv` in the current directory. You're not ready to activate your virtual environment. Run this into the command line:

```
source venv/bin/activate
```

### Install the dev build of `formulaic` package
Note: You must already be setup to use ssh with Mozilla repos in Github*. 

Install the dev package as hosted here on Github:

```
pip install git+ssh://git@github.com/Mozilla-Ocho/formulaic-python.git
```

*_If you're not setup for SSH see [Setup Git to use SSH keys](https://docs.github.com/en/authentication/connecting-to-github-with-ssh/adding-a-new-ssh-key-to-your-github-account?platform=mac&tool=webui) and [Generate SSH keys locally](https://docs.github.com/en/authentication/connecting-to-github-with-ssh/generating-a-new-ssh-key-and-adding-it-to-the-ssh-agent#generating-a-new-ssh-key)_

### Alternately: download the files and run directly
You can also download this repo to your local machine and run all of the files inside `/src/formulaic/` to take a quick spin. **Note: the tutorial below is written for installing via `pip`.** 

If you download and run all of the files inside [/src/formulaic/](https://github.com/Mozilla-Ocho/formulaic-python/tree/main/src/formulaic), use the [example.py](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/src/formulaic/example.py) file located there as your script file. This file only works if it is in the same directory as the other three files: `motivator.json`, `formulaic.py`, and `openclient.py`. 


## Quick Start
We're going to build [this script](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/quickstart.py) step-by-step below, using a [Formula JSON file](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/motivator.json) we downloaded from [Formulaic.app](http://formulaic.app). If you download both this script and the JSON file to your working directory, you can run them right away. You will need a  llamafile server running at `localhost:8080`. You can also substitute in an  OpenAI key and get going that way. We're going to break the entire thing down step-by-step in a moment. 

```
from formulaic_ai import Formula, load_formula, OpenClient
 


model_config = { "llamafile" :  {"url" : "http://localhost:8080/v1",
                                 "key":"sk-no-key-required", 
                                 "name": "LLaMA_CPP"}, 
                  "OpenAI" :    {"url" :  "https://api.openai.com/v1",
                                 "key": "OPENAI_KEY_GOES_HERE", 
                                 "name": "gpt-3.5-turbo"},
                  "mistral7b" : {"url" : "https://api.endpoints.anyscale.com/v1",
                                 "key": "ANYSCALE_KEY_GOES_HERE", 
                                 "name": "mistralai/Mistral-7B-Instruct-v0.1"}  
                   
}

# load our Formula, print the Formula name: description
my_formula = Formula(load_formula("motivator.json")) 
print(f"{my_formula.name}: {my_formula.description}")


# render prompts. 
my_formula.render()
print(f"\nMy starting prompts: {my_formula.prompts}")


#our new variables here. 
data = {"occasion": "I'm scared of heights and climbing a mountain", 'language': 'German'}


# render and print our prompts
my_formula.render(data)
print(f"\nMy new prompts: {my_formula.prompts}")


# Create an OpenClient instance for llamafile
with OpenClient(my_formula, model_config["llamafile"]) as client:

    # start our chat. True = print to terminal
    client.chat(True)
    
    # print our message log.  
    print(client.messages)
```

## Step-by-step

### Do our imports

Import `Formula` which is what we'll use to work with our Formulas, `OpenClient` a wrapper on the OpenAI library to make it seamless to send Formula prompts to any OpenAI compatible API, and a helper function `load_formula` to open Formula files. 

```
from formulaic_ai import Formula, OpenClient, load_formula
```

### Load and create our Formula instance
All Formulas on [Formulaic.app](https://formulaic.app) JSON files that you can download to your local machine. We're providing one called `motivator.json` for you to use for this tutorial, which you can [download here](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/motivator.json). 

- Run `load_formula()` with a single argument of the the filepath+filename to `motivator.json`. That opens the Formula's JSON file and loads it into a Python dictionary
- Create an instance of the `Formula` class by passing it the dictionary we just created. I combined these two steps and saved my `Formula` instance as `my_formula`
- Now let's print the Formula name and description.
  
```
# load our Formula
my_formula = Formula(load_formula("motivator.json")) 

print(f"{my_formula.name}: {my_formula.description}")

```
We see:

```
Daily Motivator: Generates a motivational slogan based on your occasion.
```

### Render prompts
Our Formula is loaded correctly. Now let's call the `.render()` method. Downloaded Formula prompts often contain templating variables. When we render, we replace the template variables with values and generate prompts that are stored in the `.prompts` property. If we don't pass new values to `.render()`, it will render prompts using the Formula's default values. Render and then print again. 

```
# render prompts. 
my_formula.render()
print(f"\nMy starting prompts: {my_formula.prompts}")
```
Printed in the terminal we see we see: 

```
My starting prompts: ['You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of first day of a new job', 'Now translate that slogan into French ']
```
Our prompts are in a Python list. The occasion is "first day of a new job" and the "French". 

Now let's pass in new data, re-render our prompts, and print again.
```
#our new variables here. 
data = {"occasion": "I'm scared of heights and climbing a mountain", 'language': 'German'}

# render and print our prompts
my_formula.render(data)
print(f"\nMy new prompts: {my_formula.prompts}")
```
Now we see our prompt list, available at `.prompts` contains the new occasion and new translation language.

```
My new prompts: ["You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of I'm scared of heights and climbing a mountain", 'Now translate that slogan into German']
```

### Setup our model endpoint configuration
We have prompts that are ready to be sent off to a language model. I'm going to use llamafile for this tutorial. [llamafile](https://github.com/Mozilla-Ocho/llamafile) is free, runs on your local machine, and easy to get going to run a local REST endpoint. I used the mistral 7B instruct llamafile. To get it running, download the file (5GB) and run it from the command line to start the local REST server. Please see the full [llamafile documentation](https://github.com/Mozilla-Ocho/llamafile). 


I went ahead and created a `model_config` dictionary to hold my model config variables to make it simpler. Beyond llamafile, I added placeholders for OpenAI and Anyscale. We can use the Formulaic Library to send our prompts to any language model API that supports the OpenAI format, so I included OpenAI and Anyscale. Anyscale provides hosting for many open source language models with an OpenAI compatible endpoint. You would have to create keys for OpenAI and Anyscale and substitute them in below. 

```

model_config = { "llamafile" :  {"url" : "http://localhost:8080/v1",
                                 "key":"sk-no-key-required", 
                                 "name": "LLaMA_CPP"}, 
                  "OpenAI" :    {"url" :  "https://api.openai.com/v1",
                                 "key": "OPENAI_KEY_GOES_HERE", 
                                 "name": "gpt-3.5-turbo"},
                  "mistral7b" : {"url" : "https://api.endpoints.anyscale.com/v1",
                                 "key": "ANYSCALE_KEY_GOES_HERE", 
                                 "name": "mistralai/Mistral-7B-Instruct-v0.1"}  
                   
}
```

### Create our OpenClient and start our chat
Now we're ready to create our `OpenClient` instance, which is a class that extends `OpenAI`
- We call `OpenClass` and pass two arguments:
-- The first is our Formula, `my_formula`.
-- The second is a dictionary that contains valid values for the `url`, `key`, and `name` of the model endpoing we're going to use. In this case, we pass it the `llamafile` dictionary from our `model_config`.

We're going to call it using a `with` statement so that OpenClient's content manager will clean up for us:

```
with OpenClient(my_formula, model_config["llamafile"]) as client:

```

We now have two options. We can just iterate over the 2 prompts we have in our Formula and await their responses. We do that by calling `.run()`. Instead, we are going to have an ongoing chat by calling `.chat()`. Both `.run` and `.chat` have a single optional argument to print out all user propmts and assistant responses to terminal. The default is `False` but we are using the command line to iteract, so we pass `True`

```
client.chat(True)
```

And we're also going to add `print(client.messages)` so that we can see the full list of all messages we sent to the model and the model sent back. Our whole block looks like this:

```
# Create an OpenClient instance for llamafile
with OpenClient(my_formula, model_config["llamafile"]) as client:

    # start our chat. True = print to terminal
    client.chat(True)
    
    # print our message log.  
    print(client.messages)

```

## Save and run the script

We save it as `quickstart.py` and run it in the terminal

```
python quickstart.py

```



It takes a moment because we're running on our local hardware using llamafile. Here's what we see:

```
User: You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of I'm scared of heights and climbing a mountain

Assistant: Absolutely, I understand that fear of heights can be a significant challenge. But remember, every mountain peak is within your reach if you believe in yourself and take it one step at a time. Here's a motivating slogan for you:

"Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!"

User: Now translate that slogan into German

Assistant: Of course! The German translation of "Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!" would be:

"Berge Innerhalb von Dir besiegen: Deine Angst ist nur ein Stufenstein zu neuen Gipfeln!"

> 
```

Notice that we have iterated over both of our two prompts and received two answers from the llamafile model. The cursor is awaiting our input. Let's tell it to translate to Latin and hit Return. 

```
> Now translate to Latin
Assistant: In Latin, the phrase could be:

"Montes Intus Vincere: Timor Tuum Nec Nisi Gradus Ad Novos Culmines!"

> 
```
We see the Latin translation from the local llamafile model, and the cursor aways our next chat input. To stop the chat, just hit Return without entering any input and the loop exits. 


### See the message log printed
Our Formula instance saved every message we sent to the model and every message the assistant sent back. This is what we accessed above by printing `client.messages`


and now we see:
```
[{'role': 'user', 'content': "You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of I'm scared of heights and climbing a mountain"}, {'role': 'assistant', 'content': 'Absolutely, I understand that fear of heights can be a significant challenge. But remember, every mountain peak is within your reach if you believe in yourself and take it one step at a time. Here\'s a motivating slogan for you:\n\n"Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!"'}, {'role': 'user', 'content': 'Now translate that slogan into German'}, {'role': 'assistant', 'content': 'Of course! The German translation of "Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!" would be:\n\n"Berge Innerhalb von Dir besiegen: Deine Angst ist nur ein Stufenstein zu neuen Gipfeln!"'}, {'role': 'user', 'content': 'Now translate to Latin'}, {'role': 'assistant', 'content': 'In Latin, the phrase could be:\n\n"Montes Intus Vincere: Timor Tuum Nec Nisi Gradus Ad Novos Culmines!"'}]

```

That's the gist! You've parsed your first Formula and sent it off to a local language model. You can send it off to other model endpoints just as easily.  

You can see [the entire script](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/quickstart.py) we just produced here. 


            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "formulaic-ai",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.8",
    "maintainer_email": null,
    "keywords": "LLM, api, generative ai, language models, openai, prompt engineering, prompt scripts",
    "author": null,
    "author_email": "Javaun Moradi <formulaic@mozilla.com>",
    "download_url": "https://files.pythonhosted.org/packages/84/cc/1aa7b00fcde62dde04f77924745829a1566cea1e0b2391672c19f31f3fc7/formulaic_ai-0.1.0.4.tar.gz",
    "platform": null,
    "description": "# Formulaic Python Library \n_**NOTE: This is a project in active development. It is not stable and changes frequently. It is not ready for production use.**_\n\nThe Formulaic library makes it easy to use Formulas inside your generative AI applications. Formulas are open-licensed JSON scripts that contain AI prompts, template variables, and model configuration.  You can browse the library of existing Formulas for many popular language models at [Formulaic.app](https://formulaic.app). \n\n## Installation\nThis repo is currently private inside Mozilla until we release it as open source. \n\n### (Recommended) Create a virtual environment\nCreate a new directory on your local machine and open a terminal in that directory. Run this command:\n\n```\npython -m venv venv\n```\n\nNote, on your machine you may have to specify python3 if \"python\" is not setup as a shortcut to your current version:\n\n```\npython3 -m venv venv\n```\n\nThis command creates a 'venv' virtual environment under `./venv` in the current directory. You're not ready to activate your virtual environment. Run this into the command line:\n\n```\nsource venv/bin/activate\n```\n\n### Install the dev build of `formulaic` package\nNote: You must already be setup to use ssh with Mozilla repos in Github*. \n\nInstall the dev package as hosted here on Github:\n\n```\npip install git+ssh://git@github.com/Mozilla-Ocho/formulaic-python.git\n```\n\n*_If you're not setup for SSH see [Setup Git to use SSH keys](https://docs.github.com/en/authentication/connecting-to-github-with-ssh/adding-a-new-ssh-key-to-your-github-account?platform=mac&tool=webui) and [Generate SSH keys locally](https://docs.github.com/en/authentication/connecting-to-github-with-ssh/generating-a-new-ssh-key-and-adding-it-to-the-ssh-agent#generating-a-new-ssh-key)_\n\n### Alternately: download the files and run directly\nYou can also download this repo to your local machine and run all of the files inside `/src/formulaic/` to take a quick spin. **Note: the tutorial below is written for installing via `pip`.** \n\nIf you download and run all of the files inside [/src/formulaic/](https://github.com/Mozilla-Ocho/formulaic-python/tree/main/src/formulaic), use the [example.py](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/src/formulaic/example.py) file located there as your script file. This file only works if it is in the same directory as the other three files: `motivator.json`, `formulaic.py`, and `openclient.py`. \n\n\n## Quick Start\nWe're going to build [this script](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/quickstart.py) step-by-step below, using a [Formula JSON file](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/motivator.json) we downloaded from [Formulaic.app](http://formulaic.app). If you download both this script and the JSON file to your working directory, you can run them right away. You will need a  llamafile server running at `localhost:8080`. You can also substitute in an  OpenAI key and get going that way. We're going to break the entire thing down step-by-step in a moment. \n\n```\nfrom formulaic_ai import Formula, load_formula, OpenClient\n \n\n\nmodel_config = { \"llamafile\" :  {\"url\" : \"http://localhost:8080/v1\",\n                                 \"key\":\"sk-no-key-required\", \n                                 \"name\": \"LLaMA_CPP\"}, \n                  \"OpenAI\" :    {\"url\" :  \"https://api.openai.com/v1\",\n                                 \"key\": \"OPENAI_KEY_GOES_HERE\", \n                                 \"name\": \"gpt-3.5-turbo\"},\n                  \"mistral7b\" : {\"url\" : \"https://api.endpoints.anyscale.com/v1\",\n                                 \"key\": \"ANYSCALE_KEY_GOES_HERE\", \n                                 \"name\": \"mistralai/Mistral-7B-Instruct-v0.1\"}  \n                   \n}\n\n# load our Formula, print the Formula name: description\nmy_formula = Formula(load_formula(\"motivator.json\")) \nprint(f\"{my_formula.name}: {my_formula.description}\")\n\n\n# render prompts. \nmy_formula.render()\nprint(f\"\\nMy starting prompts: {my_formula.prompts}\")\n\n\n#our new variables here. \ndata = {\"occasion\": \"I'm scared of heights and climbing a mountain\", 'language': 'German'}\n\n\n# render and print our prompts\nmy_formula.render(data)\nprint(f\"\\nMy new prompts: {my_formula.prompts}\")\n\n\n# Create an OpenClient instance for llamafile\nwith OpenClient(my_formula, model_config[\"llamafile\"]) as client:\n\n    # start our chat. True = print to terminal\n    client.chat(True)\n    \n    # print our message log.  \n    print(client.messages)\n```\n\n## Step-by-step\n\n### Do our imports\n\nImport `Formula` which is what we'll use to work with our Formulas, `OpenClient` a wrapper on the OpenAI library to make it seamless to send Formula prompts to any OpenAI compatible API, and a helper function `load_formula` to open Formula files. \n\n```\nfrom formulaic_ai import Formula, OpenClient, load_formula\n```\n\n### Load and create our Formula instance\nAll Formulas on [Formulaic.app](https://formulaic.app) JSON files that you can download to your local machine. We're providing one called `motivator.json` for you to use for this tutorial, which you can [download here](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/motivator.json). \n\n- Run `load_formula()` with a single argument of the the filepath+filename to `motivator.json`. That opens the Formula's JSON file and loads it into a Python dictionary\n- Create an instance of the `Formula` class by passing it the dictionary we just created. I combined these two steps and saved my `Formula` instance as `my_formula`\n- Now let's print the Formula name and description.\n  \n```\n# load our Formula\nmy_formula = Formula(load_formula(\"motivator.json\")) \n\nprint(f\"{my_formula.name}: {my_formula.description}\")\n\n```\nWe see:\n\n```\nDaily Motivator: Generates a motivational slogan based on your occasion.\n```\n\n### Render prompts\nOur Formula is loaded correctly. Now let's call the `.render()` method. Downloaded Formula prompts often contain templating variables. When we render, we replace the template variables with values and generate prompts that are stored in the `.prompts` property. If we don't pass new values to `.render()`, it will render prompts using the Formula's default values. Render and then print again. \n\n```\n# render prompts. \nmy_formula.render()\nprint(f\"\\nMy starting prompts: {my_formula.prompts}\")\n```\nPrinted in the terminal we see we see: \n\n```\nMy starting prompts: ['You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of first day of a new job', 'Now translate that slogan into French ']\n```\nOur prompts are in a Python list. The occasion is \"first day of a new job\" and the \"French\". \n\nNow let's pass in new data, re-render our prompts, and print again.\n```\n#our new variables here. \ndata = {\"occasion\": \"I'm scared of heights and climbing a mountain\", 'language': 'German'}\n\n# render and print our prompts\nmy_formula.render(data)\nprint(f\"\\nMy new prompts: {my_formula.prompts}\")\n```\nNow we see our prompt list, available at `.prompts` contains the new occasion and new translation language.\n\n```\nMy new prompts: [\"You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of I'm scared of heights and climbing a mountain\", 'Now translate that slogan into German']\n```\n\n### Setup our model endpoint configuration\nWe have prompts that are ready to be sent off to a language model. I'm going to use llamafile for this tutorial. [llamafile](https://github.com/Mozilla-Ocho/llamafile) is free, runs on your local machine, and easy to get going to run a local REST endpoint. I used the mistral 7B instruct llamafile. To get it running, download the file (5GB) and run it from the command line to start the local REST server. Please see the full [llamafile documentation](https://github.com/Mozilla-Ocho/llamafile). \n\n\nI went ahead and created a `model_config` dictionary to hold my model config variables to make it simpler. Beyond llamafile, I added placeholders for OpenAI and Anyscale. We can use the Formulaic Library to send our prompts to any language model API that supports the OpenAI format, so I included OpenAI and Anyscale. Anyscale provides hosting for many open source language models with an OpenAI compatible endpoint. You would have to create keys for OpenAI and Anyscale and substitute them in below. \n\n```\n\nmodel_config = { \"llamafile\" :  {\"url\" : \"http://localhost:8080/v1\",\n                                 \"key\":\"sk-no-key-required\", \n                                 \"name\": \"LLaMA_CPP\"}, \n                  \"OpenAI\" :    {\"url\" :  \"https://api.openai.com/v1\",\n                                 \"key\": \"OPENAI_KEY_GOES_HERE\", \n                                 \"name\": \"gpt-3.5-turbo\"},\n                  \"mistral7b\" : {\"url\" : \"https://api.endpoints.anyscale.com/v1\",\n                                 \"key\": \"ANYSCALE_KEY_GOES_HERE\", \n                                 \"name\": \"mistralai/Mistral-7B-Instruct-v0.1\"}  \n                   \n}\n```\n\n### Create our OpenClient and start our chat\nNow we're ready to create our `OpenClient` instance, which is a class that extends `OpenAI`\n- We call `OpenClass` and pass two arguments:\n-- The first is our Formula, `my_formula`.\n-- The second is a dictionary that contains valid values for the `url`, `key`, and `name` of the model endpoing we're going to use. In this case, we pass it the `llamafile` dictionary from our `model_config`.\n\nWe're going to call it using a `with` statement so that OpenClient's content manager will clean up for us:\n\n```\nwith OpenClient(my_formula, model_config[\"llamafile\"]) as client:\n\n```\n\nWe now have two options. We can just iterate over the 2 prompts we have in our Formula and await their responses. We do that by calling `.run()`. Instead, we are going to have an ongoing chat by calling `.chat()`. Both `.run` and `.chat` have a single optional argument to print out all user propmts and assistant responses to terminal. The default is `False` but we are using the command line to iteract, so we pass `True`\n\n```\nclient.chat(True)\n```\n\nAnd we're also going to add `print(client.messages)` so that we can see the full list of all messages we sent to the model and the model sent back. Our whole block looks like this:\n\n```\n# Create an OpenClient instance for llamafile\nwith OpenClient(my_formula, model_config[\"llamafile\"]) as client:\n\n    # start our chat. True = print to terminal\n    client.chat(True)\n    \n    # print our message log.  \n    print(client.messages)\n\n```\n\n## Save and run the script\n\nWe save it as `quickstart.py` and run it in the terminal\n\n```\npython quickstart.py\n\n```\n\n\n\nIt takes a moment because we're running on our local hardware using llamafile. Here's what we see:\n\n```\nUser: You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of I'm scared of heights and climbing a mountain\n\nAssistant: Absolutely, I understand that fear of heights can be a significant challenge. But remember, every mountain peak is within your reach if you believe in yourself and take it one step at a time. Here's a motivating slogan for you:\n\n\"Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!\"\n\nUser: Now translate that slogan into German\n\nAssistant: Of course! The German translation of \"Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!\" would be:\n\n\"Berge Innerhalb von Dir besiegen: Deine Angst ist nur ein Stufenstein zu neuen Gipfeln!\"\n\n> \n```\n\nNotice that we have iterated over both of our two prompts and received two answers from the llamafile model. The cursor is awaiting our input. Let's tell it to translate to Latin and hit Return. \n\n```\n> Now translate to Latin\nAssistant: In Latin, the phrase could be:\n\n\"Montes Intus Vincere: Timor Tuum Nec Nisi Gradus Ad Novos Culmines!\"\n\n> \n```\nWe see the Latin translation from the local llamafile model, and the cursor aways our next chat input. To stop the chat, just hit Return without entering any input and the loop exits. \n\n\n### See the message log printed\nOur Formula instance saved every message we sent to the model and every message the assistant sent back. This is what we accessed above by printing `client.messages`\n\n\nand now we see:\n```\n[{'role': 'user', 'content': \"You are a personal motivator assistant who is direct and believes that everyone can be their best. Generate a motivating slogan for the occasion of I'm scared of heights and climbing a mountain\"}, {'role': 'assistant', 'content': 'Absolutely, I understand that fear of heights can be a significant challenge. But remember, every mountain peak is within your reach if you believe in yourself and take it one step at a time. Here\\'s a motivating slogan for you:\\n\\n\"Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!\"'}, {'role': 'user', 'content': 'Now translate that slogan into German'}, {'role': 'assistant', 'content': 'Of course! The German translation of \"Conquer the Mountain Within: Your Fear is Just a Stepping Stone to New Heights!\" would be:\\n\\n\"Berge Innerhalb von Dir besiegen: Deine Angst ist nur ein Stufenstein zu neuen Gipfeln!\"'}, {'role': 'user', 'content': 'Now translate to Latin'}, {'role': 'assistant', 'content': 'In Latin, the phrase could be:\\n\\n\"Montes Intus Vincere: Timor Tuum Nec Nisi Gradus Ad Novos Culmines!\"'}]\n\n```\n\nThat's the gist! You've parsed your first Formula and sent it off to a local language model. You can send it off to other model endpoints just as easily.  \n\nYou can see [the entire script](https://github.com/Mozilla-Ocho/formulaic-python/blob/main/examples/quickstart.py) we just produced here. \n\n",
    "bugtrack_url": null,
    "license": null,
    "summary": "The official Formulaic python library for working with generative AI prompts.",
    "version": "0.1.0.4",
    "project_urls": {
        "documentation": "https://docs.formulaic.app",
        "homepage": "https:formulaic.app",
        "repository": "https://github.com/Mozilla-Ocho/formulaic-python"
    },
    "split_keywords": [
        "llm",
        " api",
        " generative ai",
        " language models",
        " openai",
        " prompt engineering",
        " prompt scripts"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "d4c2d3a5ba2f168748278eb6d36899d9d1139e0c2ef12bf187c31d3e5541be54",
                "md5": "4c4e38da49730894f9b6b446cb23df68",
                "sha256": "2348f95cd1d7c3e2015128846db6a2f269701312838a35ba12227d62542f6320"
            },
            "downloads": -1,
            "filename": "formulaic_ai-0.1.0.4-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "4c4e38da49730894f9b6b446cb23df68",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.8",
            "size": 14169,
            "upload_time": "2024-03-25T22:02:21",
            "upload_time_iso_8601": "2024-03-25T22:02:21.171538Z",
            "url": "https://files.pythonhosted.org/packages/d4/c2/d3a5ba2f168748278eb6d36899d9d1139e0c2ef12bf187c31d3e5541be54/formulaic_ai-0.1.0.4-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "84cc1aa7b00fcde62dde04f77924745829a1566cea1e0b2391672c19f31f3fc7",
                "md5": "8d174930db75e12352827e59adad77e7",
                "sha256": "f3b1130858edef03a4e00f2f79a75dba1d35d5e9de116a9fc6ccf18a677966d6"
            },
            "downloads": -1,
            "filename": "formulaic_ai-0.1.0.4.tar.gz",
            "has_sig": false,
            "md5_digest": "8d174930db75e12352827e59adad77e7",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.8",
            "size": 14513,
            "upload_time": "2024-03-25T22:02:22",
            "upload_time_iso_8601": "2024-03-25T22:02:22.533070Z",
            "url": "https://files.pythonhosted.org/packages/84/cc/1aa7b00fcde62dde04f77924745829a1566cea1e0b2391672c19f31f3fc7/formulaic_ai-0.1.0.4.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-03-25 22:02:22",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "Mozilla-Ocho",
    "github_project": "formulaic-python",
    "github_not_found": true,
    "lcname": "formulaic-ai"
}
        
Elapsed time: 0.20716s