Instantly scale AI and machine learning workloads on GPU on-demand
Functionality Technology
September 17, 2024 / September 17, 2024 by UbiOps
In a linkedin post by Fergal Mcgovern in May, he tries to explain why around 83% of enterprise CIOs plan to place some workloads on-premise instead of on-cloud. Let’s briefly explain what we mean when we say on-cloud and on-premise: On-cloud storage is when data is stored in data centers operated by third parties which […]
Read more »
August 29, 2024 / August 29, 2024 by UbiOps
Machine learning operations (MLOps) involve a set of techniques and principles aimed at the design, development, deployment, and maintenance of machine learning models for production use. The purpose of MLOps is to establish a clear set of guidelines to simplify the complex process of bringing a model into production. You can also learn more about […]
Tagged
Deploy your model Functionality
August 23, 2024 / August 23, 2024 by UbiOps
In recent times we’ve seen that open-source LLMs like Mixtral and Llama are starting to rival the performance of some proprietary LLMs. One of the things to consider when working with open-source models though is that they do not come ready to go for every use case out of the box, like the lack of […]
Functionality Review product
July 12, 2024 / July 12, 2024 by UbiOps
What UbiOps delivers more than standard Model Serving Platforms? Model serving is the process of providing access to production-level models for end-users or applications. Meaning that they will be deployed for internal or external usage . In most cases, such as with UbiOps, they will be available via a REST API. This stage is very […]
Functionality
June 21, 2024 / June 21, 2024 by UbiOps
Multi-model routing is a process of linking multiple AI models together. The routing can either be done in series or in parallel, meaning that you use a router to send prompts to specific models. Example of a simple multi-model route Multi-modal routing can have various sorts of benefits. It enables you to have smaller and […]
Functionality LLM
May 28, 2024 / May 28, 2024 by UbiOps
Reducing inference costs for GenAI
May 22, 2024 / May 22, 2024 by UbiOps
In a previous article we showed how you can set up a Retrievel Augmented Generation (RAG) framework for the Mistral-7B-v.02 Instruct LLM using the UbiOps WebApp. In this article we’ll go a step further and create a front-end for that set-up using Streamlit, and we’ll be using the UbiOps Python Client Library to set-up the […]
May 15, 2024 / May 15, 2024 by UbiOps
In this guide, we will show you how to increase data throughput for LLMs using batching, specifically by utilizing the vLLM library. We will explain some of the techniques it leverages and show why they are useful. We will be looking at the PagedAttention algorithm in particular. Our setup will achieve impressive performance results and […]
May 3, 2024 / May 3, 2024 by UbiOps
Introduction Optimizing inference is a machine learning (ML) engineer’s task. In a lot of cases, though, it tends to fall into the hands of data scientists. Whether you’re a data scientist deploying models as a hobby or whether you work in a team that lacks engineers, at some point you will probably have to start […]
March 28, 2024 / March 28, 2024 by UbiOps
In this article, we will be creating a chatbot which is fine-tuned on custom documentation. We’ll use UbiOps—which is an AI deployment, serving and management platform—to fine-tune and deploy the instruction-tuned Mistral-7B model taken from Hugging Face. We’ll explain some of the methods used to fine-tune models, such as instruction tuning and domain adaptation, but […]
Deploy your model Functionality Technology UbiOps
March 19, 2024 / March 21, 2024 by UbiOps
Model deployment or model serving designates the stage in which a trained model is brought to production and readily usable. A model-serving platform allows you to easily deploy and monitor your models hassle-free. Below is the MLOps dev cycle and how UbiOps can be used within that cycle. How UbiOps fits into the MLOps dev […]
January 30, 2024 / February 20, 2024 by UbiOps
Most of the open-source models available on Huggingface come pre-trained on a large corpus of publicly available data, like WebText. In general, the size of these datasets give large language models (LLMs) an adequate performance for various use cases. For some, more specific, use cases, however, more domain specific knowledge is required for the LLM […]
Functionality UbiOps
December 28, 2023 / July 3, 2024 by UbiOps
Facilitate a hybrid-cloud strategy and save weeks of work!
December 18, 2023 / December 18, 2023 by UbiOps
In the good old days machine learning models were made from scratch by data scientists. This involved acquiring, and cleaning data before training a model and getting it to production. In recent years, though, the size of models has increased, and thus the training data required to train these new larger models as well. This […]
August 15, 2023 / August 17, 2023 by UbiOps
Pros and cons of different techniques More and more companies are actively using artificial intelligence (AI) in their business, and, slowly but surely, more models are being brought into production. When making the step towards production, inference time starts to play an important role. When a model is external user facing, you typically want to […]
August 15, 2023 / September 6, 2023 by UbiOps
Introduction A Graphical Processing Unit (GPU) is a processor that is made up of smaller, more specialized cores. Originally designed to accelerate graphical calculations, GPUs were developed to work in parallel processing, which means that they are able to process data simultaneously in order to complete tasks more quickly. In other words, GPUs are able […]
August 4, 2023 / January 22, 2024 by UbiOps
What can you get out of this guide? In this guide, we explain how to deploy LLaMa 2, an open-source Large Language Model (LLM), using UbiOps for easy model hosting and Streamlit for creating a chatbot UI. The guide provides step-by-step instructions for packaging a deployment, loading it into UbiOps, configuring compute on GPUs and […]
July 28, 2023 / August 17, 2023 by UbiOps
Artificial Intelligence (AI) has emerged as a world-changing technology with a wide range of applications across industries. From virtual assistants to autonomous vehicles and advanced data analytics, AI has started to revolutionize the way we live and work. However, most AI algorithms require a very large amount of computational power to process and analyze all the necessary data. This is where […]
Events Functionality
UbiOps, leading platform for deploying and scaling Artificial Intelligence (AI) and Machine Learning (ML) models, is proud to introduce advanced functionality for training AI models in the cloud. This development allows businesses to manage even more of their AI development lifecycle on the UbiOps platform and also leverage Generative AI faster. Training and fine-tuning AI […]
June 25, 2023 / July 26, 2023 by UbiOps
One of the big reasons for the increased usage of AI on the web is the availability of open source foundation models. Increasingly, Artificial Intelligence (AI) lies at the heart of online tools and applications. For example, the global chatbot market is expected to reach $1 billion dollars by 2024, because they can save companies […]
Functionality Technology Whitepapers
The IT spending of financial institutions all over the world is steadily increasing, and is expected to reach over $750 billion dollars by 2025. This is partly because of the significant increase in the development and deployment of AI systems. AI-powered systems can process large volumes of data very quickly and at a large scale. […]
Collaborations Functionality Technology
June 12, 2023 / July 26, 2023 by UbiOps
Agritech & AI working together Business owners in the horticultural sector and agritech can use automated image recognition to automate their crop observations and thereby optimize harvest and work scheduling. This pioneering technology is being developed by Ridder and is in part made possible thanks to UbiOps’ computation power. The sector has the opportunity to […]
Blog Functionality Product update
June 1, 2023 / July 26, 2023 by UbiOps
Training Machine Learning models in the cloud from scratch can be a challenging task. In this post we will dive into why UbiOps is not only useful for running and scaling model inference, but can also be used to run training jobs for Machine Learning models. UbiOps has a built-in functionality for managing and running […]
Functionality Product update
June 1, 2023 / July 31, 2023 by UbiOps
UbiOps release news – version 2.24.0 On the 1st of June 2023 we released new functionality and made improvements to our UbiOps SaaS platform. On this page we will walk you through the changes with some examples of how to use the new functionality Python client library version for this release: 3.15.0CLI version for this release: 2.15.0 […]