Instantly scale AI and machine learning workloads on GPU on-demand
Deploy your model
July 31, 2024 / July 31, 2024 by UbiOps
In this guide, we’ll take you through the release of the new update from MetaAI. This update saw changes to the existing Llama 3 8B & 70B models, while also releasing a new model with 405B parameters (Llama 3.1 405B). We’ll also deploy a quantized version of the Llama 3.1 8B Instruct model to UbiOps. […]
Read more »
Tagged
Collaborations
June 20, 2024 / June 20, 2024 by UbiOps
How UbiOps and Arize help you stay in control LLMs are all the rage at the moment, and the APIs of closed source models like GPT-4 have made it easier than ever to leverage the power of AI. However, for a lot of regulated industries these closed source models are not an option. Luckily there […]
Functionality Technology
May 22, 2024 / May 22, 2024 by UbiOps
In a previous article we showed how you can set up a Retrievel Augmented Generation (RAG) framework for the Mistral-7B-v.02 Instruct LLM using the UbiOps WebApp. In this article we’ll go a step further and create a front-end for that set-up using Streamlit, and we’ll be using the UbiOps Python Client Library to set-up the […]
Functionality
May 3, 2024 / May 3, 2024 by UbiOps
Introduction Optimizing inference is a machine learning (ML) engineer’s task. In a lot of cases, though, it tends to fall into the hands of data scientists. Whether you’re a data scientist deploying models as a hobby or whether you work in a team that lacks engineers, at some point you will probably have to start […]
Deploy your model Technology
March 5, 2024 / March 11, 2024 by UbiOps
Introduction In a recent workshop on Large Language Models (LLMs), we asked attendees which LLMs they’re using or considering for their use cases. It turns out everyone was primarily focused on open-source LLMs like Mistral 7B, Llama 2, and Falcon. This left us wondering about a recent report by Reuters on OpenAI’s growth. The report […]
Technology UbiOps
January 31, 2024 / February 1, 2024 by UbiOps
Given the number of Large Language Models (LLMs) out there, finding one that meets your specific use case can be a daunting task. The field is evolving rapidly, with new models and fine-tuned versions being released every single week. It follows that any list of LLMs and how they should be applied will be rapidly […]
January 30, 2024 / February 20, 2024 by UbiOps
Most of the open-source models available on Huggingface come pre-trained on a large corpus of publicly available data, like WebText. In general, the size of these datasets give large language models (LLMs) an adequate performance for various use cases. For some, more specific, use cases, however, more domain specific knowledge is required for the LLM […]
December 18, 2023 / December 18, 2023 by UbiOps
In the good old days machine learning models were made from scratch by data scientists. This involved acquiring, and cleaning data before training a model and getting it to production. In recent years, though, the size of models has increased, and thus the training data required to train these new larger models as well. This […]