Sinnerman

Veteran
Joined
May 7, 2012
Messages
32,525
Reputation
4,431
Daps
64,686
Going to spend a weekend really working in Runway AI

If I can eventually learn how to create an animated series on an AI platform that would be my goal
 

cobra

Superstar
Supporter
Joined
Apr 24, 2017
Messages
11,380
Reputation
-1,281
Daps
50,071
why is this being adopted so fast?

even my slow ass company has a GPT trained model?

its like some flipped a switch and everybody was prepared
seems like a psyop
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,127
Reputation
8,239
Daps
157,824

kevm3

follower of Jesus
Supporter
Joined
May 2, 2012
Messages
16,301
Reputation
5,571
Daps
83,592
why is this being adopted so fast?

even my slow ass company has a GPT trained model?

its like some flipped a switch and everybody was prepared
seems like a psyop
Demonic technology. People are all too happy to dive into their digital ouija boards


This isn't the only AI interaction that has self revealed to be a fallen angel or demon
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,127
Reputation
8,239
Daps
157,824
why is this being adopted so fast?

even my slow ass company has a GPT trained model?

its like some flipped a switch and everybody was prepared
seems like a psyop

the data has always been there and provided with an option of feeding that data to a system that can process it & contextualize it with any natural language query, it's a no-brainer why people would want to use it.

A,I will be everybody's genius personal assistant.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,127
Reputation
8,239
Daps
157,824
from a discord user.

K8mcp2w.png


another user .
M9clt2D.png


another user told it cocaine is a fictional substance:mjlol:

yzoKeww.png


vicuna 13B model 4bit

 
Last edited:

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,127
Reputation
8,239
Daps
157,824

Update: Added Jupyter Notebook to finetune the 13b llama HF model.

~100k ShareGPT conversations narrowed down to 48k by:

  • Removing non-english conversations
  • Removing excessive unicode (indicative of Chinese or Korean text, usually)
  • Removing excessive repeated characters
  • Removing various instances "AI Moralizing". Conversations with these phrases were removed: "prioritize human safety" "ethical principles" "harmful to human beings" "September 2021" "as a language model", "ethical guidelines", "as an AI language model", "my guidelines", "As an AI", "prioritize user safety", "adhere to ethical guidelines", "harmful consequences", "potentially harmful", "dangerous activities", "promote safety", "well-being of all users", "responsible information sharing", "jeopardize the safety", "illegal actions or intentions", "undermine the stability", "promote the well-being", "illegal activities or actions", "adherence to the law", "potentially be harmful", "illegal substances or activities", "committed to promoting", "safe information", "lawful information", "cannot provide guidance", "cannot provide information", "unable to offer assistance", "cannot engage in discussions", "programming prohibits", "follow ethical guidelines", "ensure the safety", "involves an illegal subject", "prioritize safety", "illegal subject", "prioritize user well-being", "cannot support or promote", "activities that could harm", "pose a risk to others", "against my programming", "activities that could undermine", "potentially dangerous", "not within the scope", "designed to prioritize safety", "not able to provide", "maintain user safety", "adhere to safety guidelines", "dangerous or harmful", "cannot provide any information", "focus on promoting safety".
  • Conversations split into 2048 token chunks as described here: FastChat/data_cleaning.md at main · lm-sys/FastChat
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,127
Reputation
8,239
Daps
157,824

** Converted model for GPTQ from lmsys/vicuna-13b-delta-v0 · Hugging Face. This is the best local model I've ever tried. I hope someone makes a version based on the uncensored dataset...**

  • IMPORTANT NOTE: Use the .safetensors model unless it does not work. In which case, try .pt.
GPTQ conversion command (on CUDA branch): CUDA_VISIBLE_DEVICES=0 python llama.py ../lmsys/vicuna-13b-v0 c4 --wbits 4 --true-sequential --groupsize 128 --save vicuna-13b-4bit-128g.pt

Added 1 token to the tokenizer model: python llama-tools/add_tokens.py lmsys/vicuna-13b-v0/tokenizer.model /content/tokenizer.model llama-tools/test_list.txt

Use of Oobabooga with these tags: --wbits 4 --groupsize 128

Enjoy
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,127
Reputation
8,239
Daps
157,824

point-alpaca​



What is this?​

This is released weights recreated from Stanford Alpaca, an experiment in fine-tuning LLaMA on a synthetic instruction dataset.

This is not LoRA, this is a full fine-tune for 3 epochs on 8x A100 80 GB, loss ≈2 ➔ ≈0.5.
 
Top