gintonic modular blockchain for AI
  • Gintonic Hub
  • Live Subchains
    • Distillery Intro
    • Inference Intro
    • PrivateAI Intro
  • Planned Subchains
    • Service Discovery Intro
    • language Processing Intro
    • Forecasting Intro
    • Model Monitoring Intro
    • Deployment Intro
    • Pipeline Intro
    • AI Governance Intro
    • Training Intro
    • Fine-tuning Intro
    • Storage Intro
    • Data Transfer Intro
    • Data Lake Intro
  • Inference Subchain Guide
    • Getting Started
    • Core Concepts
    • API Reference
    • Best Practices
    • Billing and Usage
    • Troubleshooting
    • FAQs
  • Distillery Subchain Guide
    • Getting Started
    • Joining the Network
    • Earning Rewards
    • Managing Your Node
    • Troubleshooting
    • Glossary of Terms
    • FAQ
  • Private AI Subchain Guide
Powered by GitBook
On this page
  • Introduction to the inference subchain
  • What is the inference subchain?
  • Why Choose the inference subchain?
  • How Does It Work?
  • Ready to Dive In?

Inference Subchain Guide

Comprehensive guide to the Gintonic inference subchain. Learn how to leverage cost-effective, scalable AI model execution for your development projects.

Introduction to the inference subchain

Welcome to the Gintonic inference subchain - your gateway to cost-effective, scalable AI model execution. If you're an AI developer looking to leverage powerful language models without breaking the bank, you're in the right place.

What is the inference subchain?

The inference subchain is a crucial component of the Gintonic ecosystem. It's designed to bridge the gap between developers and our distributed AI infrastructure. Think of it as the API layer that lets you tap into the power of large language models (starting with Mistral LLM) without the headache of managing complex infrastructure.

Why Choose the inference subchain?

  1. Cheaper than your morning coffee (well, almost) Our distributed approach means you're not paying for idle cloud resources. Run your models at a fraction of the cost of traditional centralized services.

  2. Scale like a pro, pay like a novice Whether you're handling 100 or 1,000,000 requests, the inference subchain scales seamlessly. You only pay for what you use.

  3. Familiar territory, new possibilities If you've used services like AWS Bedrock or Google Vertex AI, you'll feel right at home. We've designed our API to be intuitive for developers coming from other platforms.

  4. Cutting-edge models, old-school pricing Start with the Mistral LLM, and stay tuned as we add more state-of-the-art models to our roster.

How Does It Work?

Here's the TL;DR:

  1. You send an inference request through our API.

  2. We route it to our distillery subchain (our secret sauce for distributed processing).

  3. A network of GPU-powered nodes crunches the numbers.

  4. You get back the results, fast and cheap.

It's distributed computing magic, and you don't need to understand the inner workings to reap the benefits.

See It in Action: AI Speakeasy

By creating an account and chatting with the AI bartender, you can experience firsthand the responsiveness and capabilities of our platform. It's a great way to envision how you might integrate similar functionality into your own applications.

Ready to Dive In?

Great! Head over to the Getting Started section to set up your account and make your first API call. If you're the type who likes to understand the nitty-gritty details first, check out our Core Concepts page.

Remember, we're here to make AI inference accessible and affordable. If you run into any issues or have questions, don't hesitate to reach out to our support team.

Let's start building smarter applications together!

PreviousData Lake IntroNextGetting Started

Last updated 8 months ago

To get a taste of what's possible with the inference subchain, check out . This interactive demo features an AI bartender powered by the Mistral model running on our inference subchain.

AI Speakeasy
Try AI Speakeasy