Fediverse ebooks bot using neural networks https://social.exozy.me/@ta180m/107833248217767420
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
Anthony Wang dc56b0c327
Fix bug where bot post isn't created with --yes flag
4 weeks ago
.gitignore Initial commit 6 months ago
LICENSE Switch to AGPL license 6 months ago
README.md Add background info to the README 6 months ago
bot.py Fix bug where bot post isn't created with --yes flag 4 weeks ago
bot_lstm.py Add new transformers bot script and move old one to bot_lstm.py 6 months ago
data.py Modify data generation script for transformers 6 months ago
dataset.py More cleanup 6 months ago
model.py More cleanup 6 months ago
predict.py Take top 5 instead of top 3 6 months ago
train.py Adjust training parameters to train gpt2-large 1 month ago
train_lstm.py Move train.py to train_lstm.py and add new transformers training code 6 months ago



Fediverse ebooks bot using neural networks


Text generation programs have existed for decades. However, most ebooks bots today rely on Markov chains, which are fast but also produce text that sounds like you pulled some random words out of a hat (which isn't entirely inaccurate in this case).

However, we have another solution. A very overhyped solution, yes. Neural networks! Here are some samples from produced by this bot:

I toot. I don't want to. I'm happy with it. I like to make people laugh. The only thing I want is to use it to do a wonderful job.

This is total BS, and there really is no difference between this and the Matrix Matrix Matrix Matrix Matrix.

Follow me for the next few days. Please remember, I'm sorry, and I'm sorry for the inconvenience.

@Gargron @gargron @mattkat @craj_chris I am not a lawyer. I am just the voice of God. I'm a non-profit organization and I can be seen in other ways.

As you can see, neural networks can generate much more coherent text and learn how to use mentions and hashtags. The caveat? It takes only a few hours to train the network, and text generation takes a few seconds, compared to Markov chains which can do that almost instaneously.

This bot consists of three components:

  • The data.py script accesses the fediverse server's database and retrieves all messages to form the training data for the neural network.

  • The train.py script downloads a pre-trained DistilGPT2 (you can use a larger model like GPT-J if your hardware is powerful enough) and fine-tunes it on the training data from the database. By using a pre-trained model, the bot will already have a wide variety of knowledge about topics, including ones not even mentioned in the training data. This step takes a long time but you only have to do it once.

  • The bot.py script uses the fine-tuned model to generate text, and posts it to your fediverse server.


First, install Python dependencies using your distro's package manager or pip: psycopg2, torch, transformers, and datasets. Additionally, for Mastodon and Pleroma, install Mastodon.py, for Misskey, install Misskey.py, and for Matrix, install simplematrixbotlib. If your database or platform isn't supported, don't worry! It's easy to add support for other platforms and databases, and contributions are welcome!

Now retrieve the training data from your fediverse server's database using python data.py -d 'dbname=test user=postgres password=secret host=localhost port=5432'. Retrieving the training data from the database is not yet supported for Matrix. You can skip this step if you have collected training data from another source.

Next, train the network with python train.py, which may take several hours. It's a lot faster when using a GPU. If you need advanced features when training, you can also train using run_clm.py.

Finally, create an application for your bot account and generate an access token. Run the bot with python bot.py -b server_type -i fediverse.instance -t access_token. You can omit -b server_type for Mastodon and Pleroma. To run the bot periodically, create a cron job. Enjoy!