HelpingAI Series
Collection
Our Emotionally intelligent Models • 7 items • Updated • 2
How to use HelpingAI/Helpingai3-raw with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("text-generation", model="HelpingAI/Helpingai3-raw")
messages = [
{"role": "user", "content": "Who are you?"},
]
pipe(messages) # Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("HelpingAI/Helpingai3-raw")
model = AutoModelForCausalLM.from_pretrained("HelpingAI/Helpingai3-raw")
messages = [
{"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))How to use HelpingAI/Helpingai3-raw with vLLM:
# Install vLLM from pip:
pip install vllm
# Start the vLLM server:
vllm serve "HelpingAI/Helpingai3-raw"
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:8000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "HelpingAI/Helpingai3-raw",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'docker model run hf.co/HelpingAI/Helpingai3-raw
How to use HelpingAI/Helpingai3-raw with SGLang:
# Install SGLang from pip:
pip install sglang
# Start the SGLang server:
python3 -m sglang.launch_server \
--model-path "HelpingAI/Helpingai3-raw" \
--host 0.0.0.0 \
--port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "HelpingAI/Helpingai3-raw",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'docker run --gpus all \
--shm-size 32g \
-p 30000:30000 \
-v ~/.cache/huggingface:/root/.cache/huggingface \
--env "HF_TOKEN=<secret>" \
--ipc=host \
lmsysorg/sglang:latest \
python3 -m sglang.launch_server \
--model-path "HelpingAI/Helpingai3-raw" \
--host 0.0.0.0 \
--port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
-H "Content-Type: application/json" \
--data '{
"model": "HelpingAI/Helpingai3-raw",
"messages": [
{
"role": "user",
"content": "What is the capital of France?"
}
]
}'How to use HelpingAI/Helpingai3-raw with Docker Model Runner:
docker model run hf.co/HelpingAI/Helpingai3-raw
HelpingAI3 is an advanced language model developed to excel in emotionally intelligent conversations. Building upon the foundations of HelpingAI2.5, this model offers enhanced emotional understanding and contextual awareness.
HelpingAI3 was trained on a diverse dataset comprising:
The model underwent the following training processes:
HelpingAI3 is designed for:
While HelpingAI3 strives for high emotional intelligence, users should be aware of potential limitations:
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load the HelpingAI3 model
model = AutoModelForCausalLM.from_pretrained("HelpingAI/HAI3-RAW")
# Load the tokenizer
tokenizer = AutoTokenizer.from_pretrained("HelpingAI/HAI3-RAW")
# Define the chat input
chat = [
{"role": "system", "content": "You are HelpingAI, an emotional AI. Always answer my questions in the HelpingAI style."},
{"role": "user", "content": "Introduce yourself."}
]
inputs = tokenizer.apply_chat_template(
chat,
add_generation_prompt=True,
return_tensors="pt"
).to(model.device)
# Generate text
outputs = model.generate(
inputs,
max_new_tokens=256,
do_sample=True,
temperature=0.6,
top_p=0.9,
)
response = outputs[0][inputs.shape[-1]:]
print(tokenizer.decode(response, skip_special_tokens=True))