Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions playbooks/backup/dify-ai-agents/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Build Node Based AI Agents and RAG Workflows with Dify

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/dify-ai-agents/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "dify-ai-agents",
"title": "Build Node Based AI Agents and RAG Workflows with Dify",
"description": "Create AI agents and RAG workflows using Dify's visual node editor with llama.cpp on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["dify", "agents", "rag", "llamacpp", "workflows"]
}
3 changes: 3 additions & 0 deletions playbooks/backup/gguf-quantization-export/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Quantize and Export Models to GGUF

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/gguf-quantization-export/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "gguf-quantization-export",
"title": "Quantize and Export Models to GGUF",
"description": "Learn how to quantize and export models to GGUF format using llama.cpp on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["gguf", "quantization", "llamacpp", "model-export"]
}
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
# Fine-tuning on vLLM
# Local Foundry

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/local-foundry/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "local-foundry",
"title": "Local Foundry",
"description": "Set up and use Local Foundry on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": true,
"isFeatured": false,
"published": true,
"tags": ["foundry", "local", "llm"]
}
3 changes: 3 additions & 0 deletions playbooks/backup/mcp-research-agent/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Building a Research Agent Using MCP

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/mcp-research-agent/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "mcp-research-agent",
"title": "Building a Research Agent Using MCP",
"description": "Build a research agent using the Model Context Protocol (MCP) on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": true,
"isFeatured": false,
"published": true,
"tags": ["mcp", "agents", "research", "llm"]
}
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
# Creating Multi-agent Chatbots
# Getting Started with Ollama

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/ollama-getting-started/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "ollama-getting-started",
"title": "Getting Started with Ollama",
"description": "Install and run LLMs locally using Ollama on your STX Halo™",
"time": 30,
"platforms": ["linux", "windows"],
"difficulty": "beginner",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["ollama", "llm", "inference", "local"]
}
3 changes: 3 additions & 0 deletions playbooks/backup/open-webui-chat/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# How to Chat with LLMs in Open WebUI

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/open-webui-chat/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "open-webui-chat",
"title": "How to Chat with LLMs in Open WebUI",
"description": "Set up Open WebUI to chat with local LLMs using llama.cpp on your STX Halo™",
"time": 30,
"platforms": ["linux", "windows"],
"difficulty": "beginner",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["openwebui", "chat", "llamacpp", "llm"]
}
3 changes: 3 additions & 0 deletions playbooks/backup/openhands-agent-sdk/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# OpenHands Software Agent SDK

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/openhands-agent-sdk/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "openhands-agent-sdk",
"title": "OpenHands Software Agent SDK",
"description": "Build software agents using the OpenHands Agent SDK on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": true,
"isFeatured": false,
"published": true,
"tags": ["openhands", "agents", "sdk", "llm"]
}
3 changes: 3 additions & 0 deletions playbooks/backup/openhands-coding-assistant/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# How to Use Local LLMs as a Coding Assistant in OpenHands

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/openhands-coding-assistant/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "openhands-coding-assistant",
"title": "How to Use Local LLMs as a Coding Assistant in OpenHands",
"description": "Configure OpenHands to use local LLMs as a coding assistant with llama.cpp on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["openhands", "coding-assistant", "llamacpp", "qwen"]
}
3 changes: 3 additions & 0 deletions playbooks/backup/tailscale-remote-access/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Setting up Tailscale To Remotely Access Halo

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/backup/tailscale-remote-access/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "tailscale-remote-access",
"title": "Setting up Tailscale To Remotely Access Halo",
"description": "Configure Tailscale for secure remote access to your STX Halo™ from anywhere",
"time": 30,
"platforms": ["linux", "windows"],
"difficulty": "beginner",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["tailscale", "remote-access", "networking", "vpn"]
}
12 changes: 0 additions & 12 deletions playbooks/core/vllm-inference/playbook.json

This file was deleted.

Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
{
"id": "3dgs-pytorch-rocm",
"id": "3dgs-pytorch",
"title": "3D Gaussian Splatting with PyTorch ROCm",
"description": "Render stunning 3D scenes using Gaussian Splatting with PyTorch and ROCm on STX Halo™",
"time": 90,
Expand Down
3 changes: 3 additions & 0 deletions playbooks/supplemental/clustering-rccl/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Clustering with Two Halos (RCCL)

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/supplemental/clustering-rccl/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "clustering-rccl",
"title": "Clustering with Two Halos (RCCL)",
"description": "Set up a multi-node cluster using two STX Halo™ devices with RCCL for distributed workloads",
"time": 90,
"platforms": ["linux", "windows"],
"difficulty": "advanced",
"isNew": true,
"isFeatured": false,
"published": true,
"tags": ["clustering", "distributed", "multi-gpu", "rccl", "vllm"]
}
3 changes: 3 additions & 0 deletions playbooks/supplemental/clustering-rpc-server/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Clustering with Two Halos (RPC Server)

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/supplemental/clustering-rpc-server/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "clustering-rpc-server",
"title": "Clustering with Two Halos (RPC Server)",
"description": "Set up distributed inference using RPC server across two STX Halo™ devices with vLLM",
"time": 90,
"platforms": ["linux"],
"difficulty": "advanced",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["clustering", "rpc", "vllm", "distributed", "docker"]
}
12 changes: 0 additions & 12 deletions playbooks/supplemental/clustering-two-halos/playbook.json

This file was deleted.

3 changes: 3 additions & 0 deletions playbooks/supplemental/cvml/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Local Perception Capability with CVML SDK

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/supplemental/cvml/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "cvml",
"title": "Local Perception Capability with CVML SDK",
"description": "Build local perception capabilities using CVML SDK on top of RyzenAI and ROCm",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": true,
"isFeatured": false,
"published": true,
"tags": ["cvml", "perception", "ryzenai", "rocm"]
}
3 changes: 0 additions & 3 deletions playbooks/supplemental/dify-ai-agents/README.md

This file was deleted.

12 changes: 0 additions & 12 deletions playbooks/supplemental/dify-ai-agents/playbook.json

This file was deleted.

Original file line number Diff line number Diff line change
@@ -1,4 +1,3 @@
# DreamBooth Fine-tuning with LoRA

<!-- Playbook content goes here -->

Original file line number Diff line number Diff line change
@@ -1,13 +1,12 @@
{
"id": "dreambooth-lora-finetuning",
"title": "DreamBooth Fine-tuning with LoRA",
"description": "Fine-tune image generation models using DreamBooth and LoRA on STX Halo™",
"time": 120,
"platforms": ["windows", "linux"],
"description": "Fine-tune image generation models using DreamBooth and LoRA techniques on your STX Halo™",
"time": 90,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": false,
"isFeatured": false,
"published": false,
"tags": ["dreambooth", "lora", "fine-tuning", "image-generation", "stable-diffusion"]
"published": true,
"tags": ["dreambooth", "lora", "fine-tuning", "image-generation"]
}

3 changes: 3 additions & 0 deletions playbooks/supplemental/gaia-agents/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Getting Started Creating Agents with GAIA

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/supplemental/gaia-agents/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "gaia-agents",
"title": "Getting Started Creating Agents with GAIA",
"description": "Build and deploy AI agents using the GAIA framework with llama.cpp on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": false,
"isFeatured": false,
"published": true,
"tags": ["gaia", "agents", "llamacpp", "llm"]
}
3 changes: 0 additions & 3 deletions playbooks/supplemental/gguf-quantization-export/README.md

This file was deleted.

12 changes: 0 additions & 12 deletions playbooks/supplemental/gguf-quantization-export/playbook.json

This file was deleted.

3 changes: 3 additions & 0 deletions playbooks/supplemental/lemonade-getting-started/README.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,3 @@
# Using Lemonade Across CPU, GPU, and NPU

<!-- Playbook content goes here -->
12 changes: 12 additions & 0 deletions playbooks/supplemental/lemonade-getting-started/playbook.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,12 @@
{
"id": "lemonade-getting-started",
"title": "Using Lemonade Across CPU, GPU, and NPU",
"description": "Learn how to use Lemonade framework across CPU, GPU, and NPU on your STX Halo™",
"time": 60,
"platforms": ["linux", "windows"],
"difficulty": "intermediate",
"isNew": true,
"isFeatured": false,
"published": true,
"tags": ["lemonade", "cpu", "gpu", "npu", "inference"]
}
12 changes: 0 additions & 12 deletions playbooks/supplemental/multi-agent-chatbots/playbook.json

This file was deleted.

3 changes: 0 additions & 3 deletions playbooks/supplemental/open-webui-chat/README.md

This file was deleted.

12 changes: 0 additions & 12 deletions playbooks/supplemental/open-webui-chat/playbook.json

This file was deleted.

3 changes: 0 additions & 3 deletions playbooks/supplemental/openhands-coding-assistant/README.md

This file was deleted.

12 changes: 0 additions & 12 deletions playbooks/supplemental/openhands-coding-assistant/playbook.json

This file was deleted.

Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
{
"id": "pytorch-rocm-gpu-kernels",
"id": "pytorch-kernels",
"title": "Custom GPU Kernels with PyTorch ROCm",
"description": "Write and optimize custom GPU kernels using PyTorch and ROCm on STX Halo™",
"time": 120,
Expand Down
Original file line number Diff line number Diff line change
@@ -1,3 +1,3 @@
# Clustering with two Halos
# Getting Started with vLLM

<!-- Playbook content goes here -->
Loading
Loading