Skip to content

ChamathKB/llm-sdk

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

53 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

llm_lib

example workflow

Gem to invoke API calls to Huggingface and OpenAI LLMs.

Support both Huggingface and OpenAI models.

Install

gem install llm_lib

Usage examples

Replace 'API_KEY' with your actual OpenAI/Huggingface API key

OpenAI API call

apikey = ENV['OPENAI_API_KEY']
client = LlmLib::OpenAIClient.new(apikey)

prompt = "Once upon a time"
max_tokens = 100
response = client.chat_gpt_call(prompt, 
                        max_tokens)
puts response

HuggingFace API call

apikey = ENV['HUGGINGFACE_API_KEY']
client = LlmLib::HuggingfaceApiClient.new(apikey)

query = "Once upon a time"
response = client.hugging_bloom_call(query)

puts response

Supported models

Currently support following models, more models will be added in the future.

  • gpt-3.5-turbo
  • gpt-4
  • bigscience/bloom
  • tiiuae/falcon-40b-instruct
  • meta-llama/Llama-2-70b-chat-hf
  • databricks/dolly-v2-12b
  • google/flan-t5-xl

Use following example to invoke unavailable models from huggingface.

class HuggingfaceApiClient

    def initialize(api_key)
        @api_key = api_key
    end

    def hugging_dolly2_call(query, model = "databricks/dolly-v2-12b")
        response = HuggingFace.send(@api_key,
                                model,
                                query 
                                )
        response
    end
end

require 'llm_lib/huggingface'

For more visit the documentation.

Disclaimer

This is only a API wrapper for OpenAI and Huggingface models. Usage license, pricing, limitations, etc; are relevant to each model and provider, thus their guidelines and documentation must be read and followed before using.

About

Ruby package for LLMs

Resources

License

Stars

Watchers

Forks

Languages