๐Ÿ“š
Docs - Float16
homeapp
  • ๐Ÿš€GETTING STARTED
    • Introduction
    • Account
      • Dashboard
      • Profile
      • Payment
      • Workspace
      • Service Quota
    • LLM as a service
      • Quick Start
        • Set the credentials
      • Supported Model
      • Limitation
      • API Reference
    • One Click Deploy
      • Quick Start
        • Instance Detail
        • Re-generate API Key
        • Terminate Instance
      • Features
        • OpenAI Compatible
        • Long context and Auto scheduler
        • Quantization
        • Context caching
      • Limitation
      • Validated model
      • Endpoint Specification
    • Serverless GPU
      • Quick Start
        • Mode
        • Task Status
        • App Features
          • Project Detail
      • Tutorials
        • Hello World
        • Install new library
        • Prepare model weight
        • S3 Copy output from remote
        • R2 Copy output from remote
        • Direct upload and download
        • Server mode
        • LLM Dynamic Batching
        • Train and Inference MNIST
        • Etc.
      • CLI References
      • โ“FAQ
    • Playground
      • FloatChat
      • FloatPrompt
      • Quantize by Float16
  • ๐Ÿ“šUse Case
    • Q&A Bot (RAG)
    • Text-to-SQL
    • OpenAI with Rate Limit
    • OpenAI with Guardrail
    • Multiple Agents
    • Q&A Chatbots (RAG + Agents)
  • โœณ๏ธJourney
    • โœจThe Beginner's LLM Development Journey
    • ๐Ÿ“–Glossary
      • [English Version] LLM Glossary
      • [เธ เธฒเธฉเธฒเน„เธ—เธข] LLM Glossary
    • ๐Ÿง How to install node
  • Prompting
    • ๐Ÿ“šVariable
    • โ›“๏ธCondition
    • ๐Ÿ”จDemonstration
    • โŒ›Loop
    • ๐Ÿ“™Formatting
    • ๐ŸฃChat
    • ๐Ÿ”ŽTechnical term (Retrieve)
  • Privacy Policy
  • Terms & Conditions
Powered by GitBook
On this page
  1. ๐Ÿš€GETTING STARTED
  2. Serverless GPU

Tutorials

Hello WorldInstall new libraryPrepare model weightS3 Copy output from remoteR2 Copy output from remoteDirect upload and downloadServer modeLLM Dynamic BatchingTrain and Inference MNISTEtc.
PreviousProject DetailNextHello World