Creating an Uncensored Chat Assistant

Build Your Own Unrestricted AI Companion

Tired of censored responses from mainstream AI? Learn how to create your own chat assistant without content restrictions, giving you complete freedom in your conversations.

Uncensored Chat Assistant

A comprehensive guide to building your own unrestricted AI chat assistant that gives you complete conversational freedom.

120 minutes
Intermediate

Setting Up Your Uncensored AI Chat Assistant

Welcome to this step-by-step tutorial on creating your own uncensored AI chat assistant using the Aria framework. This guide is designed to be thorough, providing milestone tests at each stage to ensure a smooth setup process.

By the end of this guide, you'll have a fully functional, locally hosted AI chat assistant. Let's get started! 💻✨

What is Aria?

Aria is a local, uncensored AI entity designed to provide users with a private and customizable chat experience. Hosting Aria locally ensures that your interactions remain confidential, giving you full control over the AI's capabilities.

Prerequisites

Before diving into the setup, ensure your system meets the following requirements:

Operating System

Tested on Arch Linux with NVIDIA GPUs. Compatibility with other Linux distributions is likely, but not guaranteed.

Python

Version 3.12 installed.

NVIDIA GPU

Necessary for optimal performance. Ensure you have the latest drivers installed.

Docker

For containerized deployment (optional but recommended).

Note: This guide assumes familiarity with basic command-line operations.

Step 1: Clone the Aria Repository

Begin by cloning the Aria repository to your local machine.

bashClone Repository
git clone https://github.com/lef-fan/aria.git

Milestone Test 1:

Verify the repository is cloned successfully by navigating into the directory:

cd aria

Step 2: Choose Your Installation Method

Aria offers two primary installation methods:

Method 1: Server/Client Mode Using Docker (Recommended)

Build the Docker Image:

docker buildx build --tag ghcr.io/lef-fan/aria-server:latest .

Install Client Requirements:

pip install -r requirements_client.txt

Milestone Test 2:

Ensure Docker image is built and client dependencies are installed without errors.

Method 2: Non-Docker Installation

Install Server Requirements:

pip install -r requirements.txt

Install FlashAttention:

pip install --no-build-isolation flash-attn==2.7.4.post1

Note: Tested on Arch Linux with NVIDIA GPUs and Python 3.12. Compatibility with other setups may vary.

Milestone Test 3:

Confirm all dependencies are installed successfully.

Step 3: Launching Aria

Depending on your chosen installation method, follow the appropriate steps to launch Aria.

Non-Server/Client Mode

python main.py

Milestone Test 4:

Aria should initialize, downloading necessary models on the first run.

Server and Client Mode

On the Server Machine:

Using Docker:

docker run --net=host --gpus all --name aria-server -it ghcr.io/lef-fan/aria-server:latest

Without Docker:

python server.py

On the Client Machine:

Activate Virtual Environment:

source venv/bin/activate

Run Client:

python client.py

Note: Ensure the client's configuration points to the correct server IP address.

Milestone Test 5:

The client should connect to the server, enabling interaction with Aria.

Step 4: Configuring Aria

Customize Aria's behavior by editing the configuration file to suit your device's capabilities and use case.

Milestone Test 6:

After modifications, restart Aria to apply changes and verify functionality.

Step 5: Testing Your AI Chat Assistant

With Aria up and running, it's time to test its capabilities.

  • Initiate a Chat Session: Start a conversation to assess response quality.
  • Evaluate Performance: Monitor resource usage to ensure optimal operation.
  • Test Stability: Engage in extended interactions to verify system stability.

Milestone Test 7:

Aria should handle interactions smoothly, providing coherent and contextually appropriate responses.

Upcoming Features

The Aria project is continually evolving. Anticipated features include:

Android Client

For mobile device integration.

Raspberry Pi Client

To support lightweight, portable deployments.

Ollama Support

Enhancing AI capabilities and integrations.

Stay updated by regularly checking the Aria GitHub repository.

Conclusion

Congratulations! You've successfully set up your uncensored AI chat assistant using Aria. This locally hosted solution ensures privacy and offers a customizable platform for your AI interactions.

For further customization and updates, refer to the Aria GitHub repository.

Happy chatting! 🚀