Gen-atomic is an LLM-based code generation framework aims to support a wide range of atomic language units, from compiled code to semi-structured markups.
It is also a runtime library that enables developers generate language code blocks without needing to know how LLMs are trained, fine-tuned, evaluated, benchmarked, prompted, parsed, and integrated. It collects data from offline experiments to maximize runtime performance by deciding which models to use and how to ensemble them for a particular task, based on defined cost, time, and performance goals.
The ultimate client usage will be as follows:
code:str = AtomicRuntime("RegexVal-Stats.json","LocalModelsOnly","MaximizeAccuracy","MajorityVoting")
.Gen("RegexVal", "Generic IP address in IPv4 format, 4 blocks of numbers",
correct_cases : ["127.0.0.1", "192.1.1.12", "255.255.255.255"],
incorrect_cases: ["127:0:0:1" ,"1.0.0.0.0.0","256.1.1.999"]
)
print(code)
#prints: ^(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\.(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\.(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)\.(25[0-5]|2[0-4][0-9]|[01]?[0-9][0-9]?)$
It supports various types of plugins, such as LangUnits, Models, Prompts, Parsers, and Flows, to extend and customize its functionality.
Currently in its early alpha stage, the project is still undergoing significant changes and improvements. We will share more about the theoretical foundation and goals of the project. Some hints include Atomic Design, Flow Engineering, N-version Programming, Specification-by-example. Please refer to the Project Glossary on the Documentation page for clarification of the terms we use.
Check out the streamlit app to get an idea of what we are trying to accomplish. Sorry for the bad UX. We are new to streamlit.
-
Extract project files to a folder
-
Run the setup by typing
.\setup.sh
-
Download Ollama
- Run main method of the file 'ExperimentHost.py'. You will see the default experiment/benchmark scores like the following:
-- REGEXVAL EXPERIMENT --
+---------------+--------------------+-----------------+---------------+--------------------+------------------------+---------------------+---------------------+----------------+--------------------+
| | ol-codellama (%) | ol-llama3 (%) | ol-phi3 (%) | ol-codegemma (%) | ol-codellama:70b (%) | ol-llama3:70b (%) | ol-starcoder2 (%) | ol-gemma (%) | ol-tinyllama (%) |
|---------------+--------------------+-----------------+---------------+--------------------+------------------------+---------------------+---------------------+----------------+--------------------|
| CorrectCase | 76.74 | 48.84 | 46.51 | 65.12 | 18.60 | 79.07 | 0.00 | 16.28 | 0.00 |
| IncorrectCase | 85.48 | 95.16 | 93.55 | 83.87 | 96.77 | 74.19 | 100.00 | 95.16 | 100.00 |
| Overall | 81.90 | 76.19 | 74.29 | 76.19 | 64.76 | 76.19 | 59.05 | 62.86 | 59.05 |
+---------------+--------------------+-----------------+---------------+--------------------+------------------------+---------------------+---------------------+----------------+--------------------+
See a full experiment log from the link.
In the project root directory, run 'pytest' to test integrations:
pytest
or
.\pytest.sh
Python 3.8+, Ollama This project might depend on some certain LLM Hosts and AI frameworks in the future, such as LlamaIndex, LangChain, HF, Pytorch etc.
You can attend one of our remote meetings. We read papers and share ideas. Feel free to drop us a line.~~
Excited to have you onboard! Here's how you can contribute:
- Choose a Task. You can check out the Project tasks, or look through the open issues or suggest your own improvements via discussion board.
- Fork the Project: Make a copy of our project on your GitHub.
- Make Changes: Write your code and test it thoroughly. Ensure that pytest passes successfully.
- Submit a Pull Request: Share your changes with us for review.
Currently, datasets are version controlled. Please send a pull request for now (see 'src/data' folder). We are planning to use a data versioning system for this.