
GeForce GTX 1650 vs Radeon 890M

GeForce GTX 1650
Popular choices:

Radeon 890M
Popular choices:
Performance Spectrum - GPU
About G3D Mark
G3D Mark is a standard benchmark that measures graphics performance in real-world gaming scenarios. It simplifies comparing cards from different brands, where higher scores directly correlate with better fps and smoother gaming experiences.
Value Upgrade Path
This is the official ChipVERSUS Value Rating, comparing raw performance (G3D Mark) per dollar. Components placed above yours deliver better value for money.
Avg price is the current average price collected from markets across the web.
Performance Per Dollar
Performance Per Dollar Radeon 890M
Performance Comparison
About G3D Mark🏆 Chipversus Verdict
🚀 Performance Leadership
The Radeon 890M is the superior choice for raw performance. It leads with a 3.9% higher G3D Mark score. This advantage makes it significantly better for higher resolutions (1440p/4K) and graphic-intensive titles compared to the GeForce GTX 1650.
| Insight | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| Performance | ❌Lower raw frame rates (-3.9%) | ✅Leading raw performance (+3.9%) |
| Longevity | Turing (2018−2022) (12nm) | RDNA 3.5 (2024−2025) (4nm) |
| Ecosystem | Supports FSR Upscaling | Supports FSR Upscaling |
| VRAM | ❌ Less VRAM capacity | 🎮 High Capacity (4 GB) |
| Efficiency | ⚡ Higher Power Consumption | 💡 Excellent Perf/Watt |
| Case Fit | 📏 Compact / SFF Friendly | — |
💎 Value Proposition
While current pricing data is unavailable, the Radeon 890M remains the clear technical winner. Check real-time availability to determine if the performance gap justifies the market price.
Performance Check
Real-world benchmarks and performance projections based on comprehensive hardware analysis and comparative metrics. Values represent expected performance on High/Ultra settings at 1080p, 1440p, and 4K. Modeled using a Ryzen 7 7800X3D reference profile to minimize specific CPU bottlenecks.
Note: Performance behavior can vary per game. Specific architectures may perform better or worse depending on game engine optimizations and API implementation.
Technical Specifications
Side-by-side comparison of GeForce GTX 1650 and Radeon 890M

GeForce GTX 1650
The GeForce GTX 1650 is manufactured by NVIDIA. It was released in April 23 2019. It features the Turing architecture. The core clock ranges from 1485 MHz to 1665 MHz. It has 896 shading units. The thermal design power (TDP) is 75W. Manufactured using 12 nm process technology. G3D Mark benchmark score: 7,869 points. Launch price was $149.

Radeon 890M
The Radeon 890M is manufactured by AMD. It was released in July 15 2024. It features the RDNA 3.5 architecture. The core clock ranges from 400 MHz to 2900 MHz. It has 1024 shading units. The thermal design power (TDP) is 15W. Manufactured using 4 nm process technology. It features 16 dedicated ray tracing cores for enhanced lighting effects. G3D Mark benchmark score: 8,175 points.
Graphics Performance
The GeForce GTX 1650 scores 7,869 and the Radeon 890M reaches 8,175 in the G3D Mark benchmark — just a 3.9% difference, making them near-identical in rasterization performance. The GeForce GTX 1650 is built on Turing while the Radeon 890M uses RDNA 3.5, both on 12 nm vs 4 nm. Shader units: 896 (GeForce GTX 1650) vs 1,024 (Radeon 890M). Raw compute: 2.984 TFLOPS (GeForce GTX 1650) vs 5.939 TFLOPS (Radeon 890M). Boost clocks: 1665 MHz vs 2900 MHz.
| Feature | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| G3D Mark Score | 7,869 | 8,175+4% |
| Architecture | Turing | RDNA 3.5 |
| Process Node | 12 nm | 4 nm |
| Shading Units | 896 | 1024+14% |
| Compute (TFLOPS) | 2.984 TFLOPS | 5.939 TFLOPS+99% |
| Boost Clock | 1665 MHz | 2900 MHz+74% |
| ROPs | 32 | 32 |
| TMUs | 56 | 64+14% |
| L1 Cache | 896 KB+250% | 256 KB |
| L2 Cache | 1 MB | 2 MB+100% |
Advanced Features (DLSS/FSR)
| Feature | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| Upscaling Tech | FSR 2.1 (Compatible) | FSR 1.0 (Software) |
| Frame Generation | FSR 3 (Compatible) | Not Supported |
| Ray Reconstruction | No | No |
| Low Latency | Standard | AMD Anti-Lag |
Video Memory (VRAM)
Both cards feature 4 GB of video memory. Bus width: 128-bit vs System. L2 Cache: 1 MB (GeForce GTX 1650) vs 2 MB (Radeon 890M) — the Radeon 890M has significantly larger on-die cache to reduce VRAM reliance.
| Feature | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| VRAM Capacity | 4 GB | 4 GB |
| Memory Type | GDDR5 | Shared |
| Memory Bandwidth | 128 GB/s | System |
| Bus Width | 128-bit | System |
| L2 Cache | 1 MB | 2 MB+100% |
Display & API Support
DirectX support: 12 (GeForce GTX 1650) vs 12.2 (Radeon 890M). Vulkan: 1.4 vs 1.3. OpenGL: 4.6 vs 4.6. Maximum simultaneous displays: 3 vs 4.
| Feature | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| DirectX | 12 | 12.2+2% |
| Vulkan | 1.4+8% | 1.3 |
| OpenGL | 4.6 | 4.6 |
| Max Displays | 3 | 4+33% |
Media & Encoding
Hardware encoder: NVENC 5th gen (Volta) (GeForce GTX 1650) vs VCN 3.0 (Radeon 890M). Decoder: NVDEC 4th gen vs VCN 3.0. Supported codecs: H.264,H.265/HEVC,VP8,VP9 (GeForce GTX 1650) vs MPEG-2,H.264,HEVC,VP9,AV1 (Decode) (Radeon 890M).
| Feature | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| Encoder | NVENC 5th gen (Volta) | VCN 3.0 |
| Decoder | NVDEC 4th gen | VCN 3.0 |
| Codecs | H.264,H.265/HEVC,VP8,VP9 | MPEG-2,H.264,HEVC,VP9,AV1 (Decode) |
Power & Dimensions
The GeForce GTX 1650 draws 75W versus the Radeon 890M's 15W — a 133.3% difference. The Radeon 890M is more power-efficient. Recommended PSU: 300W (GeForce GTX 1650) vs 500W (Radeon 890M). Power connectors: None vs PCIe-powered. Card length: 229mm vs 0mm, occupying 2 vs 0 slots.
| Feature | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| TDP | 75W | 15W-80% |
| Recommended PSU | 300W-40% | 500W |
| Power Connector | None | PCIe-powered |
| Length | 229mm | 0mm |
| Height | 111mm | 0mm |
| Slots | 2 | 0-100% |
| Temp (Load) | 70°C | — |
| Perf/Watt | 104.9 | 545.0+420% |
Value Analysis
The Radeon 890M is the newer GPU (2024 vs 2019).
| Feature | GeForce GTX 1650 | Radeon 890M |
|---|---|---|
| MSRP | $149 | — |
| Avg Price (30d) | $75 | — |
| Codename | TU117 | Strix Point |
| Release | April 23 2019 | July 15 2024 |
| Ranking | #323 | #312 |
Top Performing GPUs
The most powerful gpus ranked by G3D Mark benchmark scores.












