Advertisement
Morkeleb

manual koboldcpp_rocm log with exception

Dec 25th, 2024 (edited)
37
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 2.96 KB | Software | 0 0
  1. C:\LargeLanguageModels\koboldcpp_rocm_files>python koboldcpp.py
  2. ***
  3. Welcome to KoboldCpp - Version 1.78.yr0-ROCm
  4. For command line arguments, please refer to --help
  5. ***
  6. Auto Selected HIP Backend...
  7.  
  8. Exiting by user request.
  9.  
  10. C:\LargeLanguageModels\koboldcpp_rocm_files>python koboldcpp.py
  11. ***
  12. Welcome to KoboldCpp - Version 1.78.yr0-ROCm
  13. For command line arguments, please refer to --help
  14. ***
  15. Auto Selected HIP Backend...
  16.  
  17. Auto Recommended GPU Layers: 45
  18. Attempting to use hipBLAS library for faster prompt ingestion. A compatible AMD GPU will be required.
  19. Initializing dynamic library: koboldcpp_hipblas.dll
  20. ==========
  21. Namespace(model='', model_param='C:/LargeLanguageModels/EVA-Qwen2.5-32B-v0.2-Q4_K_S.gguf', port=5001, port_param=5001, host='', launch=True, config=None, threads=7, usecublas=['normal', '0'], usevulkan=None, useclblast=None, usecpu=False, contextsize=16384, gpulayers=45, tensor_split=None, checkforupdates=False, ropeconfig=[0.0, 10000.0], blasbatchsize=512, blasthreads=7, lora=None, noshift=True, nofastforward=False, nommap=False, usemlock=False, noavx2=False, debugmode=0, onready='', benchmark=None, prompt='', promptlimit=100, multiuser=1, remotetunnel=False, highpriority=False, foreground=False, preloadstory=None, quiet=False, ssl=None, nocertify=False, mmproj=None, password=None, ignoremissing=False, chatcompletionsadapter=None, flashattention=True, quantkv=0, forceversion=0, smartcontext=False, unpack='', nomodel=False, showgui=False, skiplauncher=False, hordemodelname='', hordeworkername='', hordekey='', hordemaxctx=0, hordegenlen=0, sdmodel='', sdthreads=7, sdclamped=0, sdt5xxl='', sdclipl='', sdclipg='', sdvae='', sdvaeauto=False, sdquant=False, sdlora='', sdloramult=1.0, whispermodel='', hordeconfig=None, sdconfig=None, noblas=False)
  22. ==========
  23. Loading model: C:\LargeLanguageModels\EVA-Qwen2.5-32B-v0.2-Q4_K_S.gguf
  24.  
  25. The reported GGUF Arch is: qwen2
  26. Arch Category: 5
  27.  
  28. ---
  29. Identified as GGUF model: (ver 6)
  30. Attempting to Load...
  31. ---
  32. Using automatic RoPE scaling for GGUF. If the model has custom RoPE settings, they'll be used directly instead!
  33. It means that the RoPE values written above will be replaced by the RoPE values indicated after loading.
  34. System Info: AVX = 1 | AVX_VNNI = 0 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | AVX512_BF16 = 0 | AMX_INT8 = 0 | FMA = 1 | NEON = 0 | SVE = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | RISCV_VECT = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | SSSE3 = 1 | VSX = 0 | MATMUL_INT8 = 0 | LLAMAFILE = 0 |
  35. Traceback (most recent call last):
  36. File "C:\LargeLanguageModels\koboldcpp_rocm_files\koboldcpp.py", line 5168, in <module>
  37. main(parser.parse_args(),start_server=True)
  38. File "C:\LargeLanguageModels\koboldcpp_rocm_files\koboldcpp.py", line 4789, in main
  39. loadok = load_model(modelname)
  40. File "C:\LargeLanguageModels\koboldcpp_rocm_files\koboldcpp.py", line 925, in load_model
  41. ret = handle.load_model(inputs)
  42. OSError: exception: access violation reading 0x0000000000000000
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement