Summary of "Run YOUR own UNCENSORED AI & Use it for Hacking"

Overview

Goal: run an “uncensored” large language model (LLM) you control on a cloud VPS so it will answer any prompt (including hacking-related prompts shown in the demo).

High level approach:

Key platforms, components and concepts

Step-by-step process (high-level tutorial)

  1. Browse HuggingFace models → filter for text-generation, Olama-compatible, and “uncensored”.
  2. Choose a model/version (note quantization and required storage/RAM).
  3. Provision a cloud VPS (the video uses Hostinger) with Olama + Open Web UI preinstalled — pick a server tier with sufficient RAM/CPU.
  4. Log into Open Web UI and create an admin account.
  5. In Open Web UI → Admin panel → Settings → Manage Models → paste the model pull/install command (from the HuggingFace/Olama page) to download the model directly to the cloud server.
  6. Wait for the model to download and finalize installation on the server.
  7. In the web UI: rename/edit models, select which model to chat with, interact via text/voice/file upload, and use the code interpreter if available.
  8. Optional: download multiple models (e.g., a coder model + a reasoning model) and switch between them in the interface.

Features demonstrated in the UI

Models demonstrated (as named in the video)

Example sizes shown during selection: 6.5 GB, 11.9 GB, ~18.7 GB (Q4KM quantization).

Demo behavior shown

Resource and cost notes

Additional content / courses referenced

Main speaker / sources

Note: subtitles were auto-generated and contain some name/number inconsistencies; model names and sizes above are summarized as presented in the video.

Category ?

Technology


Share this summary


Is the summary off?

If you think the summary is inaccurate, you can reprocess it with the latest model.

Video