sunchippss / AI-Song-Of-Ice-And-Fire

Using large language models to complete A Song of Ice and Fire.

Home Page:https://liamswayne.github.io/wow

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

AI Song of Ice and Fire Logo

George R. R. Martin's popular series "A Song of Ice and Fire" completed with large language models.

Caution: The materials folder contains SPOILERS for A Song of Ice and Fire.

The final books can be viewed here:

Self-imposed rules:

  • No human writing other than the prompts.
  • No paid tools.

Tools used:

  • ChatGPT: Outline generation and expansion. Writing.
  • ChatGPT to Markdown chrome extension: Exporting chats and feeding them back in for outline expansion.
  • Keyboard Maestro (free trial): Automating the feeding of thousands of prompts into ChatGPT.
  • Python: String splitting and formatting.

I recorded a demo of Keyboard Maestro auto-prompting ChatGPT. The program is capable of using the "continue generating" button when it appears, but it isn't shown in this video. Outside of this demo, the program has a minimum time between prompts, so it doesn't breach the hourly generation limit for ChatGPT.

ChatGPT is capable of keeping track of a massive web of characters over the course of the novel. For example, you can see in this screenshot that Illyrio phases in and out of the story, doesn't appear for hundreds of pages, but is still kept in the back of the mind of the AI for when the character is needed:

Screen Shot 2023-06-18 at 3 49 15 PM

Can the writing be detected as written by an AI? Sometimes. OpenAI's official AI detection software stated that the generated text was "very unlikely" to be written by an AI (UPDATE: OpenAI's detection tool shut down). However, there are far better AI detection tools out there that have no problem identifying this text as AI generated.

Screen Shot 2023-06-19 at 8 21 37 PM

Draft 2 (the expansion of draft 1, which was itself an expansion of the outline) was so big that it was split into four chats. At some point OpenAI decided to start gradually decreasing the maximum length of chats. As a result, draft 2 was split into progressively shorter chat sessions. The four chats (1, 2, 3, 4) are accessible in the materials folder, as well as a unified version which contains the entire draft without the prompts.

One of the chats used in the third draft ran into an error message regarding the token limit that I haven't seen anywhere else:

chatGPT_token_limit

Fully understanding this message would require knowledge of the inner workings of ChatGPT. The message preceding this error was not the longest message I gave to ChatGPT, and the chat it was sent through was far from the longest chat created for the project.

Some of the chats are so long that they crash ChatGPT when generating a shared link. Below are some samples of the what the chats used to execute this project look like:

Metrics:

  • Word count: 1,273,786
  • Mean number of characters per word: 5.116
  • Prompts served: ~10000

Miscellaneous discoveries:

  • ChatGPT prohibits the use of the word "chink" in prompts but uses the word in responses. This was discovered because ChatGPT used the phrase "chink in their armor" multiple times in draft 2, which caused a "violation of content policy" warning when fed back into ChatGPT for draft 3. As of 2023/07/18, this is still an issue that needs solving.
  • The version of ChatGPT listed at the bottom of the page is not the exact version of the service you are receiving at any given time. This entire project was run on the "ChatGPT May 24 Version", but several measurable metrics changed during the time the project was running. The maximum number of prompts per chat decreased, with draft 2 chat 1 being by far the longest chat achieved. Subsequent chats ran into this limitation much earlier. The maximum number of prompts accepted per hour also seemed to decrease, although this may be due to the high usage of my account. While the behavior of ChatGPT may not have changed since May 24th, the service itself was changed.
  • The messages per hour limit seems to be tied to the number of response tokens generated. Draft 2 generated lengthy responses, so it required 120-second gaps between prompts on multiple occasions to stay below the limit. Draft 3 generated significantly shorter responses and never ran into a limit while serving prompts with 90-second gaps.

*Disclaimer: This is NOT an attempt to "replace" George R. R. Martin's work. It's an experiment to see how AI writes extensive texts. This project is not intended to push the idea that AI can replace writers.

**Disclaimer: I have made a contribution to OpenAI to offset the costs of this project.

About

Using large language models to complete A Song of Ice and Fire.

https://liamswayne.github.io/wow

License:GNU General Public License v3.0


Languages

Language:Python 100.0%