In the early 2000 s, VMware introduced the world to virtual servers that allowed IT to start more efficient use of idle server capacity. Today, Run: AI is introducing that same hypothesi to GPUs running containerized machine learning activities on Kubernetes.
This should enable data discipline units to have access to more resources than they would normally get were they simply earmarked a certain number of available GPUs. Company CEO and co-founder Omri Geller says his company believes that part of the issue in going AI projects to market is due to static aid allocation holding back data discipline teams.
” There are many times when those important and expensive computer generators are sitting idle, while at the same time, other useds that might need more compute influence since they need to run more ventures and don’t have access to resources available because they are part of a static naming ,” Geller explained.
To solve that issue of static source quotum, Run: AI came up with a solution to virtualize those GPU resources, whether on prem or in the mas, and cause IT define by plan how those resources should be divided.
” There is a need for a specific virtualization approaches for AI and actively finagled orchestration and planning of those GPU resources, while providing the visibility and control over those calculate resources to IT organizations and AI administrators ,” he said.
Run: AI built a solution to bridge this spread between the resources IT is providing to data science teams and what they require to run a generated enterprise, while still giving IT some govern over characterizing how that works.
” We actually help companies get much more out of their infrastructure, and we do it by truly abstracting the hardware from the data science, wanting you can simply extended your venture without thinking about the underlying hardware, and at any moment in time you can consume as much compute power as you need ,” he said.
While the company is still in its early stages, and the current economic situation is hitting everyone hard, Geller attends a plaza for a answer like Run: AI because it gives customers the capacity to draw the most out of existing resources, while fixing data science squads run more efficiently.
He also is taking a realistic long view when it comes to customer acquisition during this time.” These are challenging ages for everyone ,” he says.” We have plans for longer time partnerships with our clients that are not optimized for short term incomes .”
Run: AI was founded in 2018. It has raised $ 13 million, according to Geller. The busines is based in Israel with positions in the United Nation. It currently has 25 employees and a few dozen customers.
Read more: feedproxy.google.com