• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Chat with NVIDIA RTX Tech Demo

Joined
Nov 8, 2017
Messages
151 (0.06/day)
AI...

All I see here is a glorified search machine, still, I'm sorry. So now you can burn several hundred watts locally to search within a limited dataset and wonder if the answer is actually correct. Yay!

I'm not seeing the magic, honestly.
It's better to see this for what it is : a toy. Install the thing, get some fun with it, then uninstall when you are bored. I'm still dumbfounded that people really used chatGPT for "research" when it couldn't search the web.

I wouldn't be surprised that many of Nvidia ML software are just their employees having fun and exploring various uses case of the tech. A lot of the RTX stuff is starting to become standard features in various software. Rn people are just trying stuff and what see what sticks. I wouldn't get my hope up for an actual "A.I chatbot" nor do I wish for it. Something highly intelligent but without empathy is dangerous.
 
Joined
Jan 3, 2015
Messages
2,897 (0.85/day)
System Name The beast and the little runt.
Processor Ryzen 5 5600X - Ryzen 9 5950X
Motherboard ASUS ROG STRIX B550-I GAMING - ASUS ROG Crosshair VIII Dark Hero X570
Cooling Noctua NH-L9x65 SE-AM4a - NH-D15 chromax.black with IPPC Industrial 3000 RPM 120/140 MM fans.
Memory G.SKILL TRIDENT Z ROYAL GOLD/SILVER 32 GB (2 x 16 GB and 4 x 8 GB) 3600 MHz CL14-15-15-35 1.45 volts
Video Card(s) GIGABYTE RTX 4060 OC LOW PROFILE - GIGABYTE RTX 4090 GAMING OC
Storage Samsung 980 PRO 1 TB + 2 TB - Samsung 870 EVO 4 TB - 2 x WD RED PRO 16 GB + WD ULTRASTAR 22 TB
Display(s) Asus 27" TUF VG27AQL1A and a Dell 24" for dual setup
Case Phanteks Enthoo 719/LUXE 2 BLACK
Audio Device(s) Onboard on both boards
Power Supply Phanteks Revolt X 1200W
Mouse Logitech G903 Lightspeed Wireless Gaming Mouse
Keyboard Logitech G910 Orion Spectrum
Software WINDOWS 10 PRO 64 BITS on both systems
Benchmark Scores Se more about my 2 in 1 system here: kortlink.dk/2ca4x
I have one friend in the entire world and that is my GPU. So i will have a chat with my GPU...
 
Joined
Feb 20, 2020
Messages
9,340 (6.04/day)
Location
Louisiana
System Name Ghetto Rigs z490|x99|Acer 17 Nitro 7840hs/ 5600c40-2x16/ 4060/ 1tb acer stock m.2/ 4tb sn850x
Processor 10900k w/Optimus Foundation | 5930k w/Black Noctua D15
Motherboard z490 Maximus XII Apex | x99 Sabertooth
Cooling oCool D5 res-combo/280 GTX/ Optimus Foundation/ gpu water block | Blk D15
Memory Trident-Z Royal 4000c16 2x16gb | Trident-Z 3200c14 4x8gb
Video Card(s) Titan Xp-water | evga 980ti gaming-w/ air
Storage 970evo+500gb & sn850x 4tb | 860 pro 256gb | Acer m.2 1tb/ sn850x 4tb| Many2.5" sata's ssd 3.5hdd's
Display(s) 1-AOC G2460PG 24"G-Sync 144Hz/ 2nd 1-ASUS VG248QE 24"/ 3rd LG 43" series
Case D450 | Cherry Entertainment center on Test bench
Audio Device(s) Built in Realtek x2 with 2-Insignia 2.0 sound bars & 1-LG sound bar
Power Supply EVGA 1000P2 with APC AX1500 | 850P2 with CyberPower-GX1325U
Mouse Redragon 901 Perdition x3
Keyboard G710+x3
Software Win-7 pro x3 and win-10 & 11pro x3
Benchmark Scores Are in the benchmark section
Hi,
It's times like this I'm so glad I stopped jumping on new gpu's years ago and stayed using gtx hehe
Screw the "just buy rtx leather jacket man" nonsense I've not had any issues playing games I have.
 
Joined
Jun 6, 2022
Messages
622 (0.88/day)
Bleah! The grapes are sour when you can't reach them.

Recent history: Frame Generation
Before the AMD era: Bleah! Brrrrrrr! Bad! Fail!
After the copy... sorry ... AMD's answer: OoOoO! WooooW! Fantastic! Phenomenal!
 
Joined
Oct 2, 2020
Messages
682 (0.52/day)
System Name ASUS TUF F15
Processor Intel Core i5-10300H
Motherboard ASUS FX506LHB
Cooling Laptop built-in cooling lol
Memory 20GB @ 2666 Dual Channel
Video Card(s) Intel UHD & Nvidia GTX 1650 Mobile
Storage WD Black SN770 NVMe 1TB PCIe 4.0
Display(s) Laptop built-in 144 Hz FHD screen
Audio Device(s) LOGITECH 2.1-channel
Power Supply ASUS 180W PSU (from more powerful ASUS TUF DASH F15 lol)
Mouse Logitech G604
Keyboard SteelSeries Apex 7 TKL
Software Windows 10
that's interesting. Instead of getting some cheap AI speaker, get a 30-40 series card with 8 GB+ VRAM and have fun with your pricy AI bot:D:love:
 

bug

Joined
May 22, 2015
Messages
13,265 (4.04/day)
Processor Intel i5-12600k
Motherboard Asus H670 TUF
Cooling Arctic Freezer 34
Memory 2x16GB DDR4 3600 G.Skill Ripjaws V
Video Card(s) EVGA GTX 1060 SC
Storage 500GB Samsung 970 EVO, 500GB Samsung 850 EVO, 1TB Crucial MX300 and 2TB Crucial MX500
Display(s) Dell U3219Q + HP ZR24w
Case Raijintek Thetis
Audio Device(s) Audioquest Dragonfly Red :D
Power Supply Seasonic 620W M12
Mouse Logitech G502 Proteus Core
Keyboard G.Skill KM780R
Software Arch Linux + Win10
that's interesting. Instead of getting some cheap AI speaker, get a 30-40 series card with 8 GB+ VRAM and have fun with your pricy AI bot:D:love:
How about, if you have an RTX card, you don't need to spend on AI speakers? (Wth is an AI speaker anyway?)
 
Joined
Jun 29, 2023
Messages
493 (1.54/day)
Location
Spain
System Name Gungnir
Processor Ryzen 5 7600X @1.25v
Motherboard ASUS TUF B650M-PLUS WIFI
Cooling Thermalright Peerless Assasin 120 SE Black
Memory 2x16GB DDR5 CL36 5600MHz
Video Card(s) XFX RX 6800XT Merc 319 @1.1v @2600MHz clock @2140MHz vram freq. (surprisingly stable)
Storage 1TB WD SN770 | 2TB WD Blue SATA III SSD
Display(s) 1440p 165Hz VA
Case Lian Li Lancool 215
Audio Device(s) Beyerdynamic DT 770 PRO 80Ohm
Power Supply EVGA SuperNOVA 750W 80 Plus Gold
Mouse Logitech G Pro Wireless
Keyboard Keychron V6
VR HMD The bane of my existence (Oculus Quest 2)
Bleah! The grapes are sour when you can't reach them.

Recent history: Frame Generation
Before the AMD era: Bleah! Brrrrrrr! Bad! Fail!
After the copy... sorry ... AMD's answer: OoOoO! WooooW! Fantastic! Phenomenal!
If it helps: I still don't like frame gen.
This is also absolutely nothing new, and not only can you just use chatgpt or bingchat/copilot, but you could already run LLM's on AMD GPUs.
 

bug

Joined
May 22, 2015
Messages
13,265 (4.04/day)
Processor Intel i5-12600k
Motherboard Asus H670 TUF
Cooling Arctic Freezer 34
Memory 2x16GB DDR4 3600 G.Skill Ripjaws V
Video Card(s) EVGA GTX 1060 SC
Storage 500GB Samsung 970 EVO, 500GB Samsung 850 EVO, 1TB Crucial MX300 and 2TB Crucial MX500
Display(s) Dell U3219Q + HP ZR24w
Case Raijintek Thetis
Audio Device(s) Audioquest Dragonfly Red :D
Power Supply Seasonic 620W M12
Mouse Logitech G502 Proteus Core
Keyboard G.Skill KM780R
Software Arch Linux + Win10
If it helps: I still don't like frame gen.
This is also absolutely nothing new, and not only can you just use chatgpt or bingchat/copilot, but you could already run LLM's on AMD GPUs.
You can, but running a LLM is not the same as being able to feed it docs or videos as part of your questions.
This isn't meant to be something revolutionary. Just a little bonus for those that would rather tap into a LLM without sending their data to random servers.
 
Joined
Sep 17, 2014
Messages
21,037 (5.96/day)
Location
The Washing Machine
Processor i7 8700k 4.6Ghz @ 1.24V
Motherboard AsRock Fatal1ty K6 Z370
Cooling beQuiet! Dark Rock Pro 3
Memory 16GB Corsair Vengeance LPX 3200/C16
Video Card(s) ASRock RX7900XT Phantom Gaming
Storage Samsung 850 EVO 1TB + Samsung 830 256GB + Crucial BX100 250GB + Toshiba 1TB HDD
Display(s) Gigabyte G34QWC (3440x1440)
Case Fractal Design Define R5
Audio Device(s) Harman Kardon AVR137 + 2.1
Power Supply EVGA Supernova G2 750W
Mouse XTRFY M42
Keyboard Lenovo Thinkpad Trackpoint II
Software W10 x64
Bleah! The grapes are sour when you can't reach them.

Recent history: Frame Generation
Before the AMD era: Bleah! Brrrrrrr! Bad! Fail!
After the copy... sorry ... AMD's answer: OoOoO! WooooW! Fantastic! Phenomenal!
You really oughta learn the difference between DLSS3 and FG on AMD, to help you better understand the world I think.

DLSS3 is limited to a single gen of cards.
FG can be modded in on any recent card and any game.

By extension, DLSS3 is a way to sell cards and force you to buy new cards and/or wait for Nvidia's game ready bullshit. You only get it if they allow you to.
FG is merely there as an incentive.
 
Joined
Aug 15, 2010
Messages
123 (0.02/day)
funny how NVIDIA's website says OS requirement is Windows 11, but here says Windows 10 or 11. Can you run it on W10?
 
Joined
Apr 19, 2018
Messages
1,051 (0.47/day)
Processor AMD Ryzen 9 5950X
Motherboard Asus ROG Crosshair VIII Hero WiFi
Cooling Arctic Liquid Freezer II 420
Memory 32Gb G-Skill Trident Z Neo @3806MHz C14
Video Card(s) MSI GeForce RTX2070
Storage Seagate FireCuda 530 1TB
Display(s) Samsung G9 49" Curved Ultrawide
Case Cooler Master Cosmos
Audio Device(s) O2 USB Headphone AMP
Power Supply Corsair HX850i
Mouse Logitech G502
Keyboard Cherry MX
Software Windows 11
that's interesting. Instead of getting some cheap AI speaker, get a 30-40 series card with 8 GB+ VRAM and have fun with your pricy AI bot:D:love:
8GB is not enough for real A.I. 16GB is where it starts to get good.
 
Joined
Jan 18, 2020
Messages
691 (0.44/day)
LLMs are super confident bullshit artists with narrow real world use case potential.

No idea how such a limited technology has got so overhyped.
 
Joined
Dec 22, 2011
Messages
3,890 (0.86/day)
Processor AMD Ryzen 7 3700X
Motherboard MSI MAG B550 TOMAHAWK
Cooling AMD Wraith Prism
Memory Team Group Dark Pro 8Pack Edition 3600Mhz CL16
Video Card(s) NVIDIA GeForce RTX 3080 FE
Storage Kingston A2000 1TB + Seagate HDD workhorse
Display(s) Samsung 50" QN94A Neo QLED
Case Antec 1200
Power Supply Seasonic Focus GX-850
Mouse Razer Deathadder Chroma
Keyboard Logitech UltraX
Software Windows 11
A.I. is just the latest tech PR buzzword to sell more product. It offers little value, and it certainly offers very little in the way of intelligence, as you simply can't trust the output at all.

And love the block on the RTX20x0 series. Bloody nGreedia. :banghead:

I'm surprised @cvaldes hasn't berated you for getting online and mocking the relentless march of technology.
 
Joined
Feb 10, 2023
Messages
63 (0.14/day)
System Name Desktop + SteamDeck OLED
Processor i7-10700K @ -0.130mV offset
Motherboard MSI MPG Z490 Gaming Edge WiFi
Cooling be quiet! Pure Loop 280mm
Memory Corsair Vengeance RGB Pro 32GB 3200MT/s CL16
Video Card(s) RTX 3080 Ti Founders Edition @ 1830MHz, 800mV
I told you RT was made for AI, no one listened. DLSS is just an excuse for the hardware. All you RTX believers have given Nvidia a stranglehold on the AI market and the whole world is going to suffer. Good job. :banghead:
But, the RT Cores responsible for Ray Tracing and Tensor Cores responsible machine learning are different hardware
 

LegatusLabienus

New Member
Joined
Feb 16, 2024
Messages
3 (0.03/day)
Thank god that it only analyze your Youtube usage and not what one wach on CornSub....
Why do you think this is a bad thing? It's a locally run software

Assume you could submit your... user data... in some text based format to the local LLM, it could analyse it, and offer recommendations on how to make your fantasies a reality through a thorough statistical analysis
 
Joined
Nov 11, 2016
Messages
3,123 (1.14/day)
System Name The de-ploughminator Mk-II
Processor i7 13700KF
Motherboard MSI Z790 Carbon
Cooling ID-Cooling SE-226-XT + Phanteks T30
Memory 2x16GB G.Skill DDR5 7200Cas34
Video Card(s) Asus RTX4090 TUF
Storage Kingston KC3000 2TB NVME
Display(s) LG OLED CX48"
Case Corsair 5000D Air
Audio Device(s) KEF LSX II LT speakers + KEF KC62 Subwoofer
Power Supply Corsair HX850
Mouse Razor Viper Ultimate
Keyboard Corsair K75
Software win11
I imagine this chatbot could shorten development time significantly for future RPGs.

Now every NPC can tell you the entire lore of the game LOL
 

wolf

Performance Enthusiast
Joined
May 7, 2007
Messages
7,789 (1.25/day)
System Name MightyX
Processor Ryzen 5800X3D
Motherboard Gigabyte X570 I Aorus Pro WiFi
Cooling Scythe Fuma 2
Memory 32GB DDR4 3600 CL16
Video Card(s) Asus TUF RTX3080 Deshrouded
Storage WD Black SN850X 2TB
Display(s) LG 42C2 4K OLED
Case Coolermaster NR200P
Audio Device(s) LG SN5Y / Focal Clear
Power Supply Corsair SF750 Platinum
Mouse Corsair Dark Core RBG Pro SE
Keyboard Glorious GMMK Compact w/pudding
VR HMD Meta Quest 3
Software case populated with Artic P12's
Benchmark Scores 4k120 OLED Gsync bliss
A shame I had to pass so many salty comments to read the interesting / constructive ones, there's a few broken records here.

I see a lot of potential in quickly extracting meaningful honed down data from any local datasets added, in fact I'll be proposing to my workplace adding an RTX GPU to a decent tower we already posses, feeding it hundreds of gigabytes of policies, reports, communications etc etc and seeing how far these legs can stretch.
 

W1zzard

Administrator
Staff member
Joined
May 14, 2004
Messages
27,094 (3.71/day)
Processor Ryzen 7 5700X
Memory 48 GB
Video Card(s) RTX 4080
Storage 2x HDD RAID 1, 3x M.2 NVMe
Display(s) 30" 2560x1600 + 19" 1280x1024
Software Windows 10 64-bit
funny how NVIDIA's website says OS requirement is Windows 11, but here says Windows 10 or 11. Can you run it on W10?
I see no reason why you couldn't run it on Windows 10, it's just a ton of Python stuff with an EXE GUI sitting in front of it. It's all open-source, so you can compile it on Windows XP, too, if you build the dependencies yourself, too
 

bug

Joined
May 22, 2015
Messages
13,265 (4.04/day)
Processor Intel i5-12600k
Motherboard Asus H670 TUF
Cooling Arctic Freezer 34
Memory 2x16GB DDR4 3600 G.Skill Ripjaws V
Video Card(s) EVGA GTX 1060 SC
Storage 500GB Samsung 970 EVO, 500GB Samsung 850 EVO, 1TB Crucial MX300 and 2TB Crucial MX500
Display(s) Dell U3219Q + HP ZR24w
Case Raijintek Thetis
Audio Device(s) Audioquest Dragonfly Red :D
Power Supply Seasonic 620W M12
Mouse Logitech G502 Proteus Core
Keyboard G.Skill KM780R
Software Arch Linux + Win10
I see no reason why you couldn't run it on Windows 10, it's just a ton of Python stuff with an EXE GUI sitting in front of it. It's all open-source, so you can compile it on Windows XP, too, if you build the dependencies yourself, too
Do you need to compile it if it's Python?
 
Joined
Oct 26, 2018
Messages
204 (0.10/day)
Processor Intel i5-13600KF
Motherboard ASRock Z790 PG Lightning
Cooling NZXT Kraken 240
Memory Corsair Vengeance DDR5 6400
Video Card(s) XFX RX 7800 XT
Storage Samsung 990 Pro 2 TB + Samsung 860 EVO 1TB
Display(s) Dell S2721DGF 165Hz
Case Fractal Meshify C
Power Supply Seasonic Focus 750
Mouse Logitech G502 HERO
Keyboard Logitech G512
This Nvidia developer forum has more info and resources for Chat with RTX.
There are other models that can be used(Llama 2 7B, and Code Llama 13B) plus I heard they are going to update with new Google Gemma models.

These guys made a nice guide to setting up Chat with RTX to run over LAN or WAN, as well as fix a text manipulation bug that was breaking cookies.

Chat with RTX requirements :
GeForce RTX 30 Series GPU or higher with a minimum 8GB of VRAM
Windows 10 or 11, and the latest NVIDIA GPU drivers.
 
Joined
Feb 21, 2024
Messages
4 (0.05/day)
System Name HP TP01-0066
Processor Ryzen 7 3700x
Motherboard 8643
Memory 24 GB 2667 MHZ DDR4
Video Card(s) Nvidia RTX 4060
Storage 1000 GB SSD + 4000 GB HDD
Display(s) 1920x1080 + 3840x2160
Power Supply 500 watt
How can I install the Llama 2 model. I have an 4060? please help?
 

LegatusLabienus

New Member
Joined
Feb 16, 2024
Messages
3 (0.03/day)
How can I install the Llama 2 model. I have an 4060? please help?
Hello,

There are many roads to getting a model to run locally on your computer. I'll do a brief write up but there are plenty of videos on the topic already.

1. Pick your UI

There exist a few frontend applications to run the LLMs
a. StableLM
b. text-generation-webui by oobabooga

2. Pick your model

Here are a few things you should know about models:
a. They come in several parameter sizes. Common ones are 7B (B as in Billion), 13B... all the way up to 70B and maybe even more! Generally more parameters means more accurate output, but at the cost of greater computational requirements.
b. Because even 7B parameter models are difficult to run without beastly hardware, there are groups and individuals who quantise the models. This reduces their computational requirements, with minimal loss in output quality. (https://huggingface.co/TheBloke)
c. When picking a model match the VRAM requirement to what's listed on the model page. Try going up or down a quantisation depending on the output quality/ performance.
d. The base models like GPT, LLaMa, etc. get modified, optimised and uploaded to huggingface by various users. You will likely see these modified versions to be more popular than the base model.

For example, since you're specifically interested in LLaMa 2, here's a list of quantised models based on LLaMa 2: https://huggingface.co/TheBloke?search_models=llama2&sort_models=downloads#models

3. Running your model

Now that you've picked your UI and Model, it's time to run it. Note that there are many sliders you can modify to tweak your output. Consult documentation or a tutorial to understand what these sliders do.

I'm abstracting a lot of steps. Here's a video you can follow that's more in depth:

Sorry if I'm off topic. I found myself to be of the appropriate skill level to respond to this type of request.
 
Joined
Feb 21, 2024
Messages
4 (0.05/day)
System Name HP TP01-0066
Processor Ryzen 7 3700x
Motherboard 8643
Memory 24 GB 2667 MHZ DDR4
Video Card(s) Nvidia RTX 4060
Storage 1000 GB SSD + 4000 GB HDD
Display(s) 1920x1080 + 3840x2160
Power Supply 500 watt
Hello,

There are many roads to getting a model to run locally on your computer. I'll do a brief write up but there are plenty of videos on the topic already.

1. Pick your UI

There exist a few frontend applications to run the LLMs
a. StableLM
b. text-generation-webui by oobabooga

2. Pick your model

Here are a few things you should know about models:
a. They come in several parameter sizes. Common ones are 7B (B as in Billion), 13B... all the way up to 70B and maybe even more! Generally more parameters means more accurate output, but at the cost of greater computational requirements.
b. Because even 7B parameter models are difficult to run without beastly hardware, there are groups and individuals who quantise the models. This reduces their computational requirements, with minimal loss in output quality. (https://huggingface.co/TheBloke)
c. When picking a model match the VRAM requirement to what's listed on the model page. Try going up or down a quantisation depending on the output quality/ performance.
d. The base models like GPT, LLaMa, etc. get modified, optimised and uploaded to huggingface by various users. You will likely see these modified versions to be more popular than the base model.

For example, since you're specifically interested in LLaMa 2, here's a list of quantised models based on LLaMa 2: https://huggingface.co/TheBloke?search_models=llama2&sort_models=downloads#models

3. Running your model

Now that you've picked your UI and Model, it's time to run it. Note that there are many sliders you can modify to tweak your output. Consult documentation or a tutorial to understand what these sliders do.

I'm abstracting a lot of steps. Here's a video you can follow that's more in depth:

Sorry if I'm off topic. I found myself to be of the appropriate skill level to respond to this type of request.
I looked into that, but it is too low, like 1 word per minute. On the techpowerup website, it says that I can install the LLama2 model even if I only have an 8gb card: https://www.techpowerup.com/review/nvidia-chat-with-rtx-tech-demo/2.html how do I do it?
 

W1zzard

Administrator
Staff member
Joined
May 14, 2004
Messages
27,094 (3.71/day)
Processor Ryzen 7 5700X
Memory 48 GB
Video Card(s) RTX 4080
Storage 2x HDD RAID 1, 3x M.2 NVMe
Display(s) 30" 2560x1600 + 19" 1280x1024
Software Windows 10 64-bit
I looked into that, but it is too low, like 1 word per minute. On the techpowerup website, it says that I can install the LLama2 model even if I only have an 8gb card: https://www.techpowerup.com/review/nvidia-chat-with-rtx-tech-demo/2.html how do I do it?
The installed package has several .cfg files, these are text files that you can open with notepad. Look through them, in one of them is an entry that defines how much VRAM is required to install Llama2, lower that value and you should be able to install it
 
Joined
Feb 21, 2024
Messages
4 (0.05/day)
System Name HP TP01-0066
Processor Ryzen 7 3700x
Motherboard 8643
Memory 24 GB 2667 MHZ DDR4
Video Card(s) Nvidia RTX 4060
Storage 1000 GB SSD + 4000 GB HDD
Display(s) 1920x1080 + 3840x2160
Power Supply 500 watt
The installed package has several .cfg files, these are text files that you can open with notepad. Look through them, in one of them is an entry that defines how much VRAM is required to install Llama2, lower that value and you should be able to install it
after it is installed or the installer package folder? where is the directory?
 
Top