Foobara::Ai

Implements a simple mental model of asking LLM services a question and getting an answer. Allows one
to ask a question and get an answer while specifying which service to use as
its implementation.

Installation

Typical stuff. Add gem "foobara-ai" to your Gemfile or .gemspec. Or if using in a local script you can also
gem install foobara-ai. You also need to do the same for some combination of foobara-open-ai-api,
foobara-anthropic-api, or foobara-ollama-api, depending on which models you wish to ask against.

Usage

You need to require whichever services you want to use before requiring foobara/ai.

> require "foobara/open_ai_api"
> require "foobara/anthropic_api"
> require "foobara/ollama_api"
> require "foobara/ai"
> result = Ask.run!(question: "What is the pH of honey?", model: "gpt-3.5-turbo")
> puts result
The pH of honey typically ranges between 3.2 and 4.5.
>

Contributing

Bug reports and pull requests are welcome on GitHub
at https://github.com/foobara/ai

License

The gem is available as open source under the terms of the MPL-2.0 License.
See LICENSE.txt for more info.