Fediverse ebooks bot using neural networks https://social.exozy.me/@ta180m/107833248217767420
Go to file
2022-07-19 10:35:47 -05:00
.gitignore Initial commit 2022-02-20 18:58:25 -06:00
bot.py Fix bug where bot post isn't created with --yes flag 2022-07-19 10:35:47 -05:00
bot_lstm.py Add new transformers bot script and move old one to bot_lstm.py 2022-02-22 16:58:19 -06:00
data.py Modify data generation script for transformers 2022-02-22 16:57:12 -06:00
dataset.py More cleanup 2022-02-21 15:20:00 -06:00
LICENSE Switch to AGPL license 2022-02-28 20:43:44 -06:00
model.py More cleanup 2022-02-21 15:20:00 -06:00
predict.py Take top 5 instead of top 3 2022-02-21 16:39:58 -06:00
README.md Add background info to the README 2022-03-01 08:23:09 -06:00
train.py Adjust training parameters to train gpt2-large 2022-07-16 20:20:34 -05:00
train_lstm.py Move train.py to train_lstm.py and add new transformers training code 2022-02-22 16:12:13 -06:00


Fediverse ebooks bot using neural networks


Text generation programs have existed for decades. However, most ebooks bots today rely on Markov chains, which are fast but also produce text that sounds like you pulled some random words out of a hat (which isn't entirely inaccurate in this case).

However, we have another solution. A very overhyped solution, yes. Neural networks! Here are some samples from produced by this bot:

I toot. I don't want to. I'm happy with it. I like to make people laugh. The only thing I want is to use it to do a wonderful job.

This is total BS, and there really is no difference between this and the Matrix Matrix Matrix Matrix Matrix.

Follow me for the next few days. Please remember, I'm sorry, and I'm sorry for the inconvenience.

@Gargron @gargron @mattkat @craj_chris I am not a lawyer. I am just the voice of God. I'm a non-profit organization and I can be seen in other ways.

As you can see, neural networks can generate much more coherent text and learn how to use mentions and hashtags. The caveat? It takes only a few hours to train the network, and text generation takes a few seconds, compared to Markov chains which can do that almost instaneously.

This bot consists of three components:

  • The data.py script accesses the fediverse server's database and retrieves all messages to form the training data for the neural network.

  • The train.py script downloads a pre-trained DistilGPT2 (you can use a larger model like GPT-J if your hardware is powerful enough) and fine-tunes it on the training data from the database. By using a pre-trained model, the bot will already have a wide variety of knowledge about topics, including ones not even mentioned in the training data. This step takes a long time but you only have to do it once.

  • The bot.py script uses the fine-tuned model to generate text, and posts it to your fediverse server.


First, install Python dependencies using your distro's package manager or pip: psycopg2, torch, transformers, and datasets. Additionally, for Mastodon and Pleroma, install Mastodon.py, for Misskey, install Misskey.py, and for Matrix, install simplematrixbotlib. If your database or platform isn't supported, don't worry! It's easy to add support for other platforms and databases, and contributions are welcome!

Now retrieve the training data from your fediverse server's database using python data.py -d 'dbname=test user=postgres password=secret host=localhost port=5432'. Retrieving the training data from the database is not yet supported for Matrix. You can skip this step if you have collected training data from another source.

Next, train the network with python train.py, which may take several hours. It's a lot faster when using a GPU. If you need advanced features when training, you can also train using run_clm.py.

Finally, create an application for your bot account and generate an access token. Run the bot with python bot.py -b server_type -i fediverse.instance -t access_token. You can omit -b server_type for Mastodon and Pleroma. To run the bot periodically, create a cron job. Enjoy!