📚
Docs - Float16
homeapp
  • 🚀GETTING STARTED
    • Introduction
    • Account
      • Dashboard
      • Profile
      • Payment
      • Workspace
      • Service Quota
    • LLM as a service
      • Quick Start
        • Set the credentials
      • Supported Model
      • Limitation
      • API Reference
    • One Click Deploy
      • Quick Start
        • Instance Detail
        • Re-generate API Key
        • Terminate Instance
      • Features
        • OpenAI Compatible
        • Long context and Auto scheduler
        • Quantization
        • Context caching
      • Limitation
      • Validated model
      • Endpoint Specification
    • Serverless GPU
      • Quick Start
        • Mode
        • Task Status
        • App Features
          • Project Detail
      • Tutorials
        • Hello World
        • Install new library
        • Prepare model weight
        • S3 Copy output from remote
        • R2 Copy output from remote
        • Direct upload and download
        • Server mode
        • LLM Dynamic Batching
        • Train and Inference MNIST
        • Etc.
      • CLI References
      • ❓FAQ
    • Playground
      • FloatChat
      • FloatPrompt
      • Quantize by Float16
  • 📚Use Case
    • Q&A Bot (RAG)
    • Text-to-SQL
    • OpenAI with Rate Limit
    • OpenAI with Guardrail
    • Multiple Agents
    • Q&A Chatbots (RAG + Agents)
  • ✳️Journey
    • ✨The Beginner's LLM Development Journey
    • 📖Glossary
      • [English Version] LLM Glossary
      • [ภาษาไทย] LLM Glossary
    • 🧠How to install node
  • Prompting
    • 📚Variable
    • ⛓️Condition
    • 🔨Demonstration
    • ⌛Loop
    • 📙Formatting
    • 🐣Chat
    • 🔎Technical term (Retrieve)
  • Privacy Policy
  • Terms & Conditions
Powered by GitBook
On this page
  • What is LLM as a service?
  • Pricing
  • Explore Use Case
  1. GETTING STARTED

LLM as a service

Seamless LLM Integration

PreviousService QuotaNextQuick Start

Last updated 2 months ago

This service is under maintenance.

What is LLM as a service?

LLM as a Service is a large language model on-demand API service designed for users who want to utilize LLMs without deploying or managing the extensive resources they require. It's also suitable for those in the decision-making stage who wish to test models.

Our service provides instant API access, allowing users to:

  1. Choose their desired LLM model

  2. Obtain our API key

  3. Implement the service immediately

Users can focus on developing their products without worrying about deployment complexities.

Pricing

We charge based on the number of tokens used, with separate rates for prompt and completion tokens. Pricing varies between models.

Explore Use Case

For more detailed pricing information, please visit .

🚀
this link

Q&A Bot

Create your chatbot

Text-to-SQL

Easily convert to SQL

LLM as a Service's Overview