# Quick Start

## Check all instances

<figure><img src="https://3153350778-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F06IkMLrrm1ptUT8449HF%2Fuploads%2FhSrSWeTYbTjUJEw5mkqb%2Fimage.png?alt=media&#x26;token=2b26f374-5c3f-4c00-8b6c-702912163a95" alt=""><figcaption><p>All instance page</p></figcaption></figure>

When you first access the One-Click Deploy service, you'll be presented with a table displaying all your instances, both active and inactive.

Every account is allocated a quota of 4 GPU cards, with no restrictions on the type of GPU. To check your current quota, simply click on the "Quota" button or navigate to the service quota settings.

{% hint style="info" %}
Learn more about service quota [here](https://docs.float16.cloud/getting-started/account/service-quota)
{% endhint %}

## Add new instance

to start new instance:

<figure><img src="https://3153350778-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F06IkMLrrm1ptUT8449HF%2Fuploads%2FcrfgpeXUKz88fxz2b26a%2Fimage.png?alt=media&#x26;token=d93adff1-30f6-4b9e-878b-a74173c8140c" alt=""><figcaption><p>Input model repository</p></figcaption></figure>

1. Paste Hugging Face model repository and token (if required), then click "Next"

<figure><img src="https://3153350778-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F06IkMLrrm1ptUT8449HF%2Fuploads%2FIlqW0NH9T77R4mfKyyIQ%2Fimage.png?alt=media&#x26;token=22135238-cfb7-487d-8142-2f9189033675" alt=""><figcaption><p>Create Instance</p></figcaption></figure>

2. Review model name and input instance name.
3. Configure instance, select region and GPU type.
4. Review pricing and instance summary.
5. Click "Start Deploy", wait for "Start instance successfully" notification
6. Redirected to the instance's deployment section.

{% hint style="info" %}
We currently use basic optimization techniques. Learn more in the [technical](https://docs.float16.cloud/getting-started/one-click-deploy/features) section. For model support limitations, check [here](https://docs.float16.cloud/getting-started/one-click-deploy/limitation).
{% endhint %}

## Quickly test API

After successful deployment, test your model using:

### Instance Chat Playground

<figure><img src="https://3153350778-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2F06IkMLrrm1ptUT8449HF%2Fuploads%2FYbLAoeo8wwC3Nw5kWW2K%2Fimage.png?alt=media&#x26;token=e63cb52b-2652-47a2-b6ac-661e23d47461" alt=""><figcaption><p>Chat Playground</p></figcaption></figure>

After successfully deploying your model, you can easily test its performance using the Instance Chat Playground. This convenient testing tool is readily accessible from your instance overview, allowing you to immediately interact with your deployed model.

* Default settings: temperature 0.5, max tokens 512
* Customize parameters, system prompt, and text message via GUI

### cURL

Or use following command:

```bash
curl -X POST http://api.float16.cloud/dedicate/JxlkeA5y2c/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer <float16-api-key>" \
  -d '{
    "model": "<your model>",
    "messages": [
      {
        "role": "system",
        "content": "You are a helpful assistant."
      },
      {
        "role": "user",
        "content": "สวัสดี"
      }
    ]
   }'
```

{% hint style="info" %}
Find copyable API formats (including OpenAI and LangChain) in the API tab.
{% endhint %}
