This repository explores innovative multimodal prompting strategies using LLama 3.1/3.2 models. It introduces new capabilities such as tokenizer customization, API stack integration, and tool calling. The ultimate goal is to leverage these technologies to create a ballet assistant tailored for specific literature and expertise.
- Highlights and Work in Progress
- Features
- Setup and Installation
- Usage
- Dependencies
- Future Enhancements
- Acknowledgments
- License
- Custom tokenizer for improved language understanding and generation in Spanish.
- Create an agent for statistical literature review and selection.
- Build a tutor for statistical concepts, supporting advanced academic workflows.
- Adapt the model to the specific needs of a ballet school and its literature.
Utilize LLama 3.1/3.2's advanced multimodal capabilities for creative applications.
Implement tool-calling strategies to enhance interactivity and functionality.
Train models on custom datasets for niche applications (e.g., ballet-specific content).
Use together.ai endpoints for model integration and API calls.
- Python 3.8 or newer.
- A package manager like
piporconda. - Access to the
Together.aiAPI.
- Clone the repository:
git clone [https://github.com/AMorQ/MultiModal_LLM.git](https://github.com/AMorQ/MultiModal_LLM.git) cd MultiModal_LLM - Create and activate the environment:
conda env create -f environment.yaml conda activate multimodal_env
- Install additional requirements (if needed):
pip install -r requirements.txt
Open the Jupyter notebooks (.ipynb) in the repository to explore multimodal prompting experiments.
Use the scripts provided to fine-tune the LLama model with your specific dataset (e.g., ballet school literature).
Experiment with the Spanish tokenizer to enhance language-specific tasks.
Test and develop the statistical literature review functionalities and other agent-based tools.
The repository relies on the following key libraries and tools:
LLama 3.1/3.2ModelsTogether.aiAPI- Python Libraries:
TensorFlowNumPyPandas
- Jupyter Notebooks
Refer to the environment.yaml or requirements.txt for a full list of dependencies.
- Expand the ballet assistant to include interactive choreography suggestions.
- Add support for additional languages in the tokenizer.
- Integrate GPT-based models for enhanced multimodal capabilities.
- Publish statistical literature review tools as standalone utilities.
Special thanks to Together.ai for providing endpoints for LLama models and enabling advanced experimentation in this project.