Please describe your proposed solution
Video Demo: GridRepublic Prototype & Proposal Summary (click image)
PROTOTYPE: Distributed inference service WebUI
SUMMARY
As our problem statement suggests, the construction and utilization of LLM inference infrastructure is costly and centralized, placing control in the hands of a few major corporations and making it inaccessible to global populations. With this project we seek to develop a decentralized, community-powered, "intelligence ecosystem" – one that operates at lower costs, and supports a wide range of languages.
The project builds upon a prototype developed under a Fund11 "Concept" grant: Wolfram: AI - LLM Distributed Inference Services. This prototype includes the following components:
-
FOR RESOURCE PROVIDERS: An easy-to-deploy LLM-Server application which can be run on computing devices with appropriate specs, and then automatically plugs into and integrates with the global inference service. "The Network is the computer", as they used to say at Sun Microsystems. (Documentation: Installing and running the GridRepublic LLM client)
-
FOR INFERENCE USERS: A range of interfaces, by which users (or 3rd party applications) can make inference requests which are fulfilled by the network of distributed resources running the LLM-Server application. Current interfaces provided by the prototype system include an API, a WebUI, and a Notebook interface. This modular system provides a large library of models, to offer users a range of price-performance options. (Demo: Distributed inference service WebUI)
The current proposal aims to further develop this prototype into a product, most notably by:
{1} Adding billing and payment systems, including ADA-based payments – to enable {a} users to pay for LLM services, and {b} resource-providers to get paid for providing the compute resources which power these services. And also,
{2} Developing the prototype into a platform for Intelligence-as-a-Service within Cardano. That is: by providing a Haskell interface and Plutus integration, we will facilitate deep integration of inference services into the Cardano ecosystem.