# co·la·la·mo
> breaking :octocat: copilot out of ⛓️ IDE ⛓️ <br/>
> so it can tell me "why is the sky blue?" <br/>
- [🧬 what I do](#-what-i-do)
- [🕹️ can I play?](#%EF%B8%8F-can-i-play)
- [install me](#install-me)
- [talk to me](#talk-to-me)
- [as a library](#as-a-library)
- [as a server](#as-a-server)
- [login](#login)
- [license](#license)
# 🧬 what I do
GitHub Copilot is a well known piece of software that primarily lives inside IDE, as a plugin, and is able to help developers with autocomplete and code snippets.
At times it feels Copilot is quite lonely to exist "as an IDE plugin" only, since it not open to be:
* called as a function
* called via an HTTP call
this is where `colalamo` comes in, because, as they say:
> _"with every **Co**pilot comes great **La**rge **La**nguage **Mo**del that makes it work"_
`colalamo` breaks Copilot free from the IDE, and makes it available:
* as a function call
* as a proxy server that accepts HTTP requests
By bringing Copilot closer to developers it opens it up to anything an LLM can do: communicating agents, retrieval augmented generation,
creating and validate user stories, explaining several source files at once, etc.
# 🕹️ can I play?
of course!
in order to use `colalamo`, you'll need a Github Copilot [subsciption](https://docs.github.com/en/billing/managing-billing-for-github-copilot/managing-your-github-copilot-individual-subscription).<br/>
it is free for open source use or/and you may have it from a company (i.e. a business subscription)
once/if you're subscribed you'll see something like this in your [settings](https://github.com/settings/billing/summary):<br/>
<img width="447" alt="image" src="https://github.com/tolitius/colalamo/assets/136575/508df008-2af6-4472-a989-6563e41b1275">
## install me
```
$ pip install colalamo
```
ready to rock! :metal:
## talk to me
`colalamo` can be used as a library or as an HTTP proxy server
### as a library
when it is used as a library:
```python
$ python
>>> from colalamo import Copilot
>>> copilot = Copilot()
>>>
>>> copilot.ask([{'content': 'how does murmur3 hash work?', 'role': 'user'}])
{'status': 200, 'text': {'reply': 'Murmur3 hash is a non-cryptographic hash function that takes an input (usually a string or binary data) and produces a fixed-size hash value as output. It was designed to be fast and efficient while providing a good distribution of hash values.\n\nHere is a simplified explanation of how Murmur3 hash works:\n\n1. Initialization: The hash function is initialized with a seed value, which is an arbitrary number chosen by the user.\n\n2. Chunking: The input data is divided into chunks of 4 bytes (32 bits) each. If the input length is not a multiple of 4, padding is added to the last chunk.\n\n3. Processing: Each chunk is processed individually. The hash function performs a series of bitwise operations, such as XOR, shift, and multiplication, on the chunk and the seed value. These operations are designed to mix the bits of the chunk and distribute them across the hash value.\n\n4. Finalization: After processing all the chunks, a finalization step is performed. It involves additional bitwise operations to further mix the bits and ensure a good distribution of the hash value.\n\n5. Output: The resulting hash value is returned as the output of the Murmur3 hash function.\n\nMurmur3 hash has several desirable properties, such as good distribution, low collision rate, and high performance. It is commonly used in applications like hash tables, bloom filters, and data indexing.', 'usage': {'completion_tokens': 286, 'prompt_tokens': 15, 'total_tokens': 301}}}
>>>
>>> ## parameters are very familiar with the ones from OpenAI API requests: top_p, temperature, n, etc.
>>> copilot.ask(messages = [{'content': 'how does murmur3 hash work?', 'role': 'user'}], temperature = 0.6)
{'status': 200, 'text': {'reply': 'MurmurHash3 is a non-cryptographic hash function that is designed to be fast and efficient while maintaining a good distribution of hash values. It was created by Austin Appleby in 2008.\n\nHere is a high-level overview of how MurmurHash3 works:\n\n1. Initialization: The hash function is initialized with a seed value that determines the output hash values.\n\n2. Chunk Processing: The input data is divided into fixed-length chunks (usually 4-byte or 8-byte chunks). These chunks are processed one at a time.\n\n3. Mixing: For each chunk, a series of bitwise operations, multiplications, and rotations are performed to mix the bits of the chunk. This mixing step helps to ensure that small changes in the input data result in significantly different hash values.\n\n4. Finalization: After all the chunks have been processed, a finalization step is performed to mix the remaining bits and produce the final hash value. This step typically involves applying additional bitwise operations and mixing the bits further.\n\n5. Output: The resulting hash value is returned as the output. It is usually a 32-bit or 64-bit integer, depending on the desired output size.\n\nMurmurHash3 is known for its speed and good distribution properties, making it suitable for a wide range of applications such as hash tables, hash-based data structures, and checksum verification. However, it is important to note that MurmurHash3 is not designed for cryptographic purposes, as it lacks the security properties required for cryptographic hash functions.', 'usage': {'completion_tokens': 307, 'prompt_tokens': 15, 'total_tokens': 322}}}
```
to check out the real use in the production code, see how <img width="42" alt="image" src="https://github.com/tolitius/colalamo/assets/136575/5b80f72d-628e-4386-813d-5c8caa231e36"> jemma [uses colalamo](https://github.com/tolitius/jemma/blob/74a770a416a7fa69a445df79baee9be50ce3e8b5/jemma/thinker.py#L206-L234)
### as a server
when it is used as a server, after it is installed all that is needed is to call it:
```bash
$ colalamo
colalamo is listening on 0.0.0.0:4242
ask away at "/ask"
example: curl http://localhost:4242/ask -X POST -d '{"messages": [{"role": "user", "content": "explain how multi-head attention work"}]}'
```
and in a different terminal / HTTP client / IDE / server, etc.
```bash
$ curl http://localhost:4242/ask -X POST -d '{"messages": [{"role": "user", "content": "explain how multi-head attention work"}]}'
{
"reply": "Multi-head attention is a key component of the Transformer model, which is widely used in natural language processing tasks such as machine translation and text summarization. The main idea behind multi-head attention is to allow the model to focus on different parts of the input sequence simultaneously, capturing various aspects of the information.\n\nHere's a step-by-step explanation of how it works:\n\n1. **Linear Projections**: The input to the multi-head attention mechanism is a set of vectors (usually the embeddings of the words in a sentence). These vectors are linearly transformed into multiple sets of Query (Q), Key (K), and Value (V) vectors. Each set is called a \"head\". The number of heads is a hyperparameter of the model.\n\n2. **Scaled Dot-Product Attention**: For each head, the model computes the attention scores by taking the dot product of the Q and K vectors, and then scaling the result by the square root of the dimension of these vectors. This is to prevent the dot product from growing too large as the dimension increases. The attention scores indicate how much each word in the sentence should be attended to.\n\n3. **Softmax Normalization**: The attention scores are then passed through a softmax function to normalize them into probabilities. This ensures that the scores are positive and sum up to 1.\n\n4. **Weighted Sum**: The softmax output is used to weight the V vectors. The weighted sum of the V vectors is the output of each head.\n\n5. **Concatenation**: The outputs of all heads are concatenated and linearly transformed to produce the final output.\n\nThe multi-head attention mechanism allows the model to capture different types of information from the input sequence. For example, one head might focus on syntactic information (e.g., the grammatical structure of the sentence), while another head might focus on semantic information (e.g., the meaning of the words).",
"usage": {
"completion_tokens": 381,
"prompt_tokens": 13,
"total_tokens": 394
}
}
```
## login
regardless whether "`colalamo`" is used as a library or a server, the first time you use it,
it would generate a code that would need to be entered on the github in order to approve this "plugin":
```python
$ python (master ✱ )
>>> from colalamo import Copilot
>>> copilot = Copilot()
don't see a token file: .copilot-token
browse to https://github.com/login/device and enter this code "4A3D-3957" to authenticate
waiting for user authorization...
```
after you go to "https://github.com/login/device" and enter the code, you'll see something similar to:
<img width="400" alt="image" src="https://github.com/tolitius/colalamo/assets/136575/eaf3cf24-ac52-43ae-b538-e4db9965c314">
colalamo will create a "`.copilot-token`" file that it will be using for all the future calls.
# license
Copyright © 2024 tolitius
Distributed under the Eclipse Public License either version 1.0 or (at
your option) any later version.
Raw data
{
"_id": null,
"home_page": null,
"name": "colalamo",
"maintainer": null,
"docs_url": null,
"requires_python": null,
"maintainer_email": null,
"keywords": "ai, copilot, github, gpt, openai, code generation",
"author": "tolitius",
"author_email": null,
"download_url": "https://files.pythonhosted.org/packages/f2/d7/7af6fdf7f854815585738a929922022023f03139efac358017421ce55f3f/colalamo-0.1.2106.tar.gz",
"platform": null,
"description": "# co\u00b7la\u00b7la\u00b7mo\n> breaking :octocat: copilot out of \u26d3\ufe0f IDE \u26d3\ufe0f <br/>\n> so it can tell me \"why is the sky blue?\" <br/>\n\n- [\ud83e\uddec what I do](#-what-i-do)\n- [\ud83d\udd79\ufe0f can I play?](#%EF%B8%8F-can-i-play)\n - [install me](#install-me)\n - [talk to me](#talk-to-me)\n - [as a library](#as-a-library)\n - [as a server](#as-a-server) \n - [login](#login)\n- [license](#license)\n\n# \ud83e\uddec what I do\n\nGitHub Copilot is a well known piece of software that primarily lives inside IDE, as a plugin, and is able to help developers with autocomplete and code snippets.\n\nAt times it feels Copilot is quite lonely to exist \"as an IDE plugin\" only, since it not open to be:\n\n* called as a function\n* called via an HTTP call\n\nthis is where `colalamo` comes in, because, as they say:\n> _\"with every **Co**pilot comes great **La**rge **La**nguage **Mo**del that makes it work\"_\n\n`colalamo` breaks Copilot free from the IDE, and makes it available:\n\n* as a function call\n* as a proxy server that accepts HTTP requests\n\nBy bringing Copilot closer to developers it opens it up to anything an LLM can do: communicating agents, retrieval augmented generation,\ncreating and validate user stories, explaining several source files at once, etc.\n\n# \ud83d\udd79\ufe0f can I play?\n\nof course!\n\nin order to use `colalamo`, you'll need a Github Copilot [subsciption](https://docs.github.com/en/billing/managing-billing-for-github-copilot/managing-your-github-copilot-individual-subscription).<br/>\nit is free for open source use or/and you may have it from a company (i.e. a business subscription)\n\nonce/if you're subscribed you'll see something like this in your [settings](https://github.com/settings/billing/summary):<br/>\n<img width=\"447\" alt=\"image\" src=\"https://github.com/tolitius/colalamo/assets/136575/508df008-2af6-4472-a989-6563e41b1275\">\n\n## install me\n\n```\n$ pip install colalamo\n```\n\nready to rock! :metal:\n\n## talk to me\n\n`colalamo` can be used as a library or as an HTTP proxy server\n\n### as a library\n\nwhen it is used as a library:\n\n```python\n$ python\n>>> from colalamo import Copilot\n>>> copilot = Copilot()\n>>>\n>>> copilot.ask([{'content': 'how does murmur3 hash work?', 'role': 'user'}])\n{'status': 200, 'text': {'reply': 'Murmur3 hash is a non-cryptographic hash function that takes an input (usually a string or binary data) and produces a fixed-size hash value as output. It was designed to be fast and efficient while providing a good distribution of hash values.\\n\\nHere is a simplified explanation of how Murmur3 hash works:\\n\\n1. Initialization: The hash function is initialized with a seed value, which is an arbitrary number chosen by the user.\\n\\n2. Chunking: The input data is divided into chunks of 4 bytes (32 bits) each. If the input length is not a multiple of 4, padding is added to the last chunk.\\n\\n3. Processing: Each chunk is processed individually. The hash function performs a series of bitwise operations, such as XOR, shift, and multiplication, on the chunk and the seed value. These operations are designed to mix the bits of the chunk and distribute them across the hash value.\\n\\n4. Finalization: After processing all the chunks, a finalization step is performed. It involves additional bitwise operations to further mix the bits and ensure a good distribution of the hash value.\\n\\n5. Output: The resulting hash value is returned as the output of the Murmur3 hash function.\\n\\nMurmur3 hash has several desirable properties, such as good distribution, low collision rate, and high performance. It is commonly used in applications like hash tables, bloom filters, and data indexing.', 'usage': {'completion_tokens': 286, 'prompt_tokens': 15, 'total_tokens': 301}}}\n>>>\n>>> ## parameters are very familiar with the ones from OpenAI API requests: top_p, temperature, n, etc.\n>>> copilot.ask(messages = [{'content': 'how does murmur3 hash work?', 'role': 'user'}], temperature = 0.6)\n{'status': 200, 'text': {'reply': 'MurmurHash3 is a non-cryptographic hash function that is designed to be fast and efficient while maintaining a good distribution of hash values. It was created by Austin Appleby in 2008.\\n\\nHere is a high-level overview of how MurmurHash3 works:\\n\\n1. Initialization: The hash function is initialized with a seed value that determines the output hash values.\\n\\n2. Chunk Processing: The input data is divided into fixed-length chunks (usually 4-byte or 8-byte chunks). These chunks are processed one at a time.\\n\\n3. Mixing: For each chunk, a series of bitwise operations, multiplications, and rotations are performed to mix the bits of the chunk. This mixing step helps to ensure that small changes in the input data result in significantly different hash values.\\n\\n4. Finalization: After all the chunks have been processed, a finalization step is performed to mix the remaining bits and produce the final hash value. This step typically involves applying additional bitwise operations and mixing the bits further.\\n\\n5. Output: The resulting hash value is returned as the output. It is usually a 32-bit or 64-bit integer, depending on the desired output size.\\n\\nMurmurHash3 is known for its speed and good distribution properties, making it suitable for a wide range of applications such as hash tables, hash-based data structures, and checksum verification. However, it is important to note that MurmurHash3 is not designed for cryptographic purposes, as it lacks the security properties required for cryptographic hash functions.', 'usage': {'completion_tokens': 307, 'prompt_tokens': 15, 'total_tokens': 322}}}\n```\n\nto check out the real use in the production code, see how <img width=\"42\" alt=\"image\" src=\"https://github.com/tolitius/colalamo/assets/136575/5b80f72d-628e-4386-813d-5c8caa231e36\"> jemma [uses colalamo](https://github.com/tolitius/jemma/blob/74a770a416a7fa69a445df79baee9be50ce3e8b5/jemma/thinker.py#L206-L234)\n\n\n\n### as a server\n\nwhen it is used as a server, after it is installed all that is needed is to call it:\n\n```bash\n$ colalamo\ncolalamo is listening on 0.0.0.0:4242\nask away at \"/ask\"\nexample: curl http://localhost:4242/ask -X POST -d '{\"messages\": [{\"role\": \"user\", \"content\": \"explain how multi-head attention work\"}]}'\n```\n\nand in a different terminal / HTTP client / IDE / server, etc.\n\n```bash\n$ curl http://localhost:4242/ask -X POST -d '{\"messages\": [{\"role\": \"user\", \"content\": \"explain how multi-head attention work\"}]}'\n{\n \"reply\": \"Multi-head attention is a key component of the Transformer model, which is widely used in natural language processing tasks such as machine translation and text summarization. The main idea behind multi-head attention is to allow the model to focus on different parts of the input sequence simultaneously, capturing various aspects of the information.\\n\\nHere's a step-by-step explanation of how it works:\\n\\n1. **Linear Projections**: The input to the multi-head attention mechanism is a set of vectors (usually the embeddings of the words in a sentence). These vectors are linearly transformed into multiple sets of Query (Q), Key (K), and Value (V) vectors. Each set is called a \\\"head\\\". The number of heads is a hyperparameter of the model.\\n\\n2. **Scaled Dot-Product Attention**: For each head, the model computes the attention scores by taking the dot product of the Q and K vectors, and then scaling the result by the square root of the dimension of these vectors. This is to prevent the dot product from growing too large as the dimension increases. The attention scores indicate how much each word in the sentence should be attended to.\\n\\n3. **Softmax Normalization**: The attention scores are then passed through a softmax function to normalize them into probabilities. This ensures that the scores are positive and sum up to 1.\\n\\n4. **Weighted Sum**: The softmax output is used to weight the V vectors. The weighted sum of the V vectors is the output of each head.\\n\\n5. **Concatenation**: The outputs of all heads are concatenated and linearly transformed to produce the final output.\\n\\nThe multi-head attention mechanism allows the model to capture different types of information from the input sequence. For example, one head might focus on syntactic information (e.g., the grammatical structure of the sentence), while another head might focus on semantic information (e.g., the meaning of the words).\",\n \"usage\": {\n \"completion_tokens\": 381,\n \"prompt_tokens\": 13,\n \"total_tokens\": 394\n }\n}\n```\n\n## login\n\nregardless whether \"`colalamo`\" is used as a library or a server, the first time you use it,\nit would generate a code that would need to be entered on the github in order to approve this \"plugin\":\n\n```python\n$ python (master \u2731 )\n\n>>> from colalamo import Copilot\n>>> copilot = Copilot()\ndon't see a token file: .copilot-token\nbrowse to https://github.com/login/device and enter this code \"4A3D-3957\" to authenticate\nwaiting for user authorization...\n```\n\nafter you go to \"https://github.com/login/device\" and enter the code, you'll see something similar to:\n\n<img width=\"400\" alt=\"image\" src=\"https://github.com/tolitius/colalamo/assets/136575/eaf3cf24-ac52-43ae-b538-e4db9965c314\">\n\ncolalamo will create a \"`.copilot-token`\" file that it will be using for all the future calls.\n\n# license\n\nCopyright \u00a9 2024 tolitius\n\nDistributed under the Eclipse Public License either version 1.0 or (at\nyour option) any later version.\n",
"bugtrack_url": null,
"license": null,
"summary": "direct api via copilot to its large language models",
"version": "0.1.2106",
"project_urls": {
"Documentation": "https://github.com/tolitius/colalamo?tab=readme-ov-file#-colalamo",
"Homepage": "https://github.com/tolitius/colalamo",
"Issues": "https://github.com/tolitius/colalamo/issues",
"Repository": "https://github.com/tolitius/colalamo"
},
"split_keywords": [
"ai",
" copilot",
" github",
" gpt",
" openai",
" code generation"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "84bb33301b90631b99516e1bcb03db1069b118d4e99f6a5d91e660942fc4939f",
"md5": "a89a10411469e659f4ad5f3faa6fca09",
"sha256": "b57b6d5d6750bb3a60b51f41b2756a09ffcb222beaacae27af1f0a122082714d"
},
"downloads": -1,
"filename": "colalamo-0.1.2106-py3-none-any.whl",
"has_sig": false,
"md5_digest": "a89a10411469e659f4ad5f3faa6fca09",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": null,
"size": 9714,
"upload_time": "2024-05-20T05:27:01",
"upload_time_iso_8601": "2024-05-20T05:27:01.614667Z",
"url": "https://files.pythonhosted.org/packages/84/bb/33301b90631b99516e1bcb03db1069b118d4e99f6a5d91e660942fc4939f/colalamo-0.1.2106-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "f2d77af6fdf7f854815585738a929922022023f03139efac358017421ce55f3f",
"md5": "df149ea8cd995030bba93eabb8978110",
"sha256": "9c8444afe546c25be016cabed9f047cc09a704ff8cc865efd0c20ca1f2d6a515"
},
"downloads": -1,
"filename": "colalamo-0.1.2106.tar.gz",
"has_sig": false,
"md5_digest": "df149ea8cd995030bba93eabb8978110",
"packagetype": "sdist",
"python_version": "source",
"requires_python": null,
"size": 12425,
"upload_time": "2024-05-20T05:27:03",
"upload_time_iso_8601": "2024-05-20T05:27:03.231857Z",
"url": "https://files.pythonhosted.org/packages/f2/d7/7af6fdf7f854815585738a929922022023f03139efac358017421ce55f3f/colalamo-0.1.2106.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-05-20 05:27:03",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "tolitius",
"github_project": "colalamo?tab=readme-ov-file#-colalamo",
"travis_ci": false,
"coveralls": false,
"github_actions": false,
"lcname": "colalamo"
}