diff --git a/Gemfile.lock b/Gemfile.lock index 0c14ccc..ce967a4 100644 --- a/Gemfile.lock +++ b/Gemfile.lock @@ -1,7 +1,7 @@ PATH remote: . specs: - ollama-ai (1.0.0) + ollama-ai (1.0.1) faraday (~> 2.9) GEM diff --git a/README.md b/README.md index b447aa9..d78451c 100644 --- a/README.md +++ b/README.md @@ -9,7 +9,7 @@ A Ruby gem for interacting with [Ollama](https://ollama.ai)'s API that allows yo ## TL;DR and Quick Start ```ruby -gem 'ollama-ai', '~> 1.0.0' +gem 'ollama-ai', '~> 1.0.1' ``` ```ruby @@ -104,11 +104,11 @@ Result: ### Installing ```sh -gem install ollama-ai -v 1.0.0 +gem install ollama-ai -v 1.0.1 ``` ```sh -gem 'ollama-ai', '~> 1.0.0' +gem 'ollama-ai', '~> 1.0.1' ``` ## Usage @@ -868,7 +868,7 @@ gem build ollama-ai.gemspec gem signin -gem push ollama-ai-1.0.0.gem +gem push ollama-ai-1.0.1.gem ``` ### Updating the README diff --git a/static/gem.rb b/static/gem.rb index 5eab5c0..6dcbc3c 100644 --- a/static/gem.rb +++ b/static/gem.rb @@ -3,7 +3,7 @@ module Ollama GEM = { name: 'ollama-ai', - version: '1.0.0', + version: '1.0.1', author: 'gbaptista', summary: 'Interact with Ollama API to run open source AI models locally.', description: "A Ruby gem for interacting with Ollama's API that allows you to run open source AI LLMs (Large Language Models) locally.", diff --git a/template.md b/template.md index 7d21e4a..413e2bf 100644 --- a/template.md +++ b/template.md @@ -9,7 +9,7 @@ A Ruby gem for interacting with [Ollama](https://ollama.ai)'s API that allows yo ## TL;DR and Quick Start ```ruby -gem 'ollama-ai', '~> 1.0.0' +gem 'ollama-ai', '~> 1.0.1' ``` ```ruby @@ -66,11 +66,11 @@ Result: ### Installing ```sh -gem install ollama-ai -v 1.0.0 +gem install ollama-ai -v 1.0.1 ``` ```sh -gem 'ollama-ai', '~> 1.0.0' +gem 'ollama-ai', '~> 1.0.1' ``` ## Usage @@ -830,7 +830,7 @@ gem build ollama-ai.gemspec gem signin -gem push ollama-ai-1.0.0.gem +gem push ollama-ai-1.0.1.gem ``` ### Updating the README