oobabooga / text-generation-webui

A Gradio web UI for Large Language Models with support for multiple inference backends.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Persistent bug with many models

elfsovereign opened this issue · comments

Describe the bug

So I've had an issue where whenever I submit a prompt, seemingly regardless of the model, it fails to generate a prompt, and can even kill the program. For reference, I'm running a RTX 2060 with 6GB of VRAM, and 16 GB of regular ram, and an intel I3-2100F processor. I know it's not the most beefcake computer ever, but I'm trying to run 7b, 3b, and 1b models (I've been trying different settings and different sizes just to see if the bug responds to lower powered models). I'm primarily looking for text generation. It seems that only GGUF models are actually working, but I don't know if that's a thing with Oobabooga overall, or a community turning away from other models, or what. I'm sadly a novice with programming and stuff, and just doing the best I can. Here's the error it prints:

 File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\gradio\queueing.py", line 527, in process_events

response = await route_utils.call_process_api(

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\gradio\route_utils.py", line 261, in call_process_api

output = await app.get_blocks().process_api(

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\gradio\blocks.py", line 1786, in process_api

result = await self.call_function(

^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\gradio\blocks.py", line 1338, in call_function

prediction = await anyio.to_thread.run_sync(

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\anyio\to_thread.py", line 56, in run_sync

return await get_async_backend().run_sync_in_worker_thread(

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\anyio\_backends\_asyncio.py", line 2505, in run_sync_in_worker_thread

return await future

^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\anyio\_backends\_asyncio.py", line 1005, in run

result = context.run(func, *args)

^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\site-packages\gradio\utils.py", line 759, in wrapper

response = f(*args, **kwargs)

^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\modules\chat.py", line 1141, in handle_character_menu_change

html = redraw_html(history, state['name1'], state['name2'], state['mode'], state['chat_style'], state['character_menu'])

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\modules\chat.py", line 490, in redraw_html

return chat_html_wrapper(history, name1, name2, mode, style, character, reset_cache=reset_cache)

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\modules\html_generator.py", line 326, in chat_html_wrapper

return generate_cai_chat_html(history['visible'], name1, name2, style, character, reset_cache)

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\modules\html_generator.py", line 250, in generate_cai_chat_html

row = [convert_to_markdown_wrapped(entry, use_cache=i != len(history) - 1) for entry in _row]

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\modules\html_generator.py", line 250, in <listcomp>

row = [convert_to_markdown_wrapped(entry, use_cache=i != len(history) - 1) for entry in _row]

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\modules\html_generator.py", line 172, in convert_to_markdown_wrapped

return convert_to_markdown.__wrapped__(string)

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\modules\html_generator.py", line 78, in convert_to_markdown

string = re.sub(pattern, replacement, string, flags=re.MULTILINE)

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

File "E:\text-generation-webui-main\installer_files\env\Lib\re\__init__.py", line 185, in sub

return _compile(pattern, flags).sub(repl, string, count)

^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

TypeError: expected string or bytes-like object, got 'NoneType' 

Is there an existing issue for this?

  • I have searched the existing issues

Reproduction

I've tried a variety of different models, and it's always an error like this (I believe it's the identical one).

Screenshot

No response

Logs

N/A

System Info

RTX 2060 with 6GB of VRAM, and 16 GB of regular ram, and an intel I3-2100F processor

Oh I rather dumbly forgot to say that I wager, given my novice-nature to advanced software like this I'd not be surprised if it was just a dumb move on my part, but I'd love to know how to fix it.

Ошибка TypeError: expected string or bytes-like object, got 'NoneType' говорит о том, что в какой-то момент код пытается обработать объект, который равен None, в функции, ожидающей строку. Эта ошибка возникает в модуле html_generator.py, где вызывается re.sub с переменной string, которая оказывается равной None.

commented

please vote my bug to, its the same
#6563