Automating your newsletters in your voice
A robot writing my newsletter

Automating your newsletters in your voice

In my last issue, I imitated a chatroom by training data from a massive Facebook Messenger chat, used Codeium to create a script to merge the data, and then loaded that into Oobabooga to create a LoRA model with Facebook’s Llama. The chat brought out everyone’s personalities though had messy formatting and was very random, so I thought of a new experiment for today:

What if I took all my newsletters from RocketFuel and trained a model off of the 40 issues? In these newsletters, I cover what happened in crypto markets and where I think things are going. In theory the model would be analytical but also be forward looking and affirmative. What if I inputted that into a LoRA model and asked it questions, would it sound like me? Well let’s find out!

I’m going to skip the step by step this time because it will be similar to my last substack (be sure to subscribe if you want to absorb my learnings!), but for reference you can see my settings here:

No alt text provided for this image
Just in case you want to do this yourself, or check the previous issue on step by step

The total text file ended being 2MB, and only took 35 minutes to process. That made me skeptical that it would affect the responses at all. However, I was wrong, it was deeply rooted in content from the newsletters. Let’s start testing this out:

If I ask it just normal questions, it seemed normal…until I started talking about investments:

No alt text provided for this image
Okay, well it already sounds like me...

Now, without the LoRA loaded in, it has no crypto bias, so here is what it said:

No alt text provided for this image

Ok let’s ask it a direct investment question:

No alt text provided for this image

I asked what I should invest in, and it chillingly said something I would probably have said even today. Which is accurate because BTC and ETH are probably also mentioned the most in the newsletters. For reference this is what the Llama model says without the model:

No alt text provided for this image
Llama 7B without my LoRA attached, no crypto

As you can see, no reference to crypto or actually any specifics. GPT4 gave a very long answer that was about diversification, risk, and random investment strategies.

What if I asked which coins to invest in?

No alt text provided for this image
Looks like this just pulled content from my newsletter, however it made up those numbers…!

As you can see it actually gave the exact format I have in my newsletters on coins that I analyze. Strangely, the numbers above are made up but very close to numbers today, I did a search for some of the market caps and it didn’t find them in the dataset. Here’s what Llama and ChatGPT said:

No alt text provided for this image
Llama 7B does not have any specifics


No alt text provided for this image
GPT4 also avoids any specifics

What if I asked it for price targets? Would it have any idea? Let’s find out:

No alt text provided for this image

Again, this chillingly sounds like my voice and even the format ending “In summary” is something I do in the newsletter. It even gives pretty sound advice with today’s prices which is very coincidental, HOWEVER, it went off the rails to talk about bots which I have never really talked about in the newsletters.

With the other models it didn’t quite work:

No alt text provided for this image

And GPT4:

No alt text provided for this image

Again a long winded answer avoiding the question, which to be fair is acceptable since the data ends in 2021.

Up until this point, we have been testing on existing content, what if we asked it to write new content in my voice?

No alt text provided for this image
Actually, it does sound like my writing style

This type of prompt isn’t good enough for me to publish, so let us be more specific to today’s news:

No alt text provided for this image
Again, it sounds like, me, but it responded like a chatbot

How about if we ask it to expand on points?

No alt text provided for this image
I have to force it to keep writing, so I need another approach

Even though some of the numbers are off, it actually sounds like it was written by me, though some of it may actually have been picked out from actual content. I needed another approach other than forcing it to keep writing and retrieving outdated data.

So far this experiment proves that you cannot automate a financial newsletter based on previous data without a lot of modifications and up to date additions. But, what if someone else wrote it, and we tried converting it to sound like me? First I’ll get a paragraph from ChatGPT:

No alt text provided for this image

I took this prompt, and asked my model to rewrite it, after some initial gibberish, it actually gave content I would have written (even if the prices were wrong):

No alt text provided for this image
Now we’re getting somewhere, even though with typos

Llama alone would just rewrite the paragraph, which was extremely similar to ChatGPT:

No alt text provided for this image
Llama 7B results just repeats it for the most part

So in theory, someone could take ChatGPT to write something long, then throw it into their own writing style LoRA model to output something that sounds like them. That doesn’t mean that the article will be accurate or avoid gibberish though. If you were to write a wide range of topics and then create a model off of it, this would work much better. It appears that if you go off topic from your content, it will just stick with Llama’s typical responses. Finally, if you have a lot more newsletters or other written content, the model will sound more and more like you with more and more topics. In theory, automating your newsletters in your voice is actually possible! It just requires a lot of general text in your style.

This was a fun experiment and I’m looking forward to seeing how others find other uses for fine tuning. Next up I will cover some stuff in the Generative Art space.

要查看或添加评论,请登录

Jeff Wang的更多文章

  • Falling into Microsoft

    Falling into Microsoft

    Taken from https://jeffwang.substack.

  • How to Hit a Homerun

    How to Hit a Homerun

    It’s your turn at bat, it’s time to take a swing, what do you do? If you’re a startup founder or VC, you’re likely…

    1 条评论
  • Race against the machine

    Race against the machine

    It’s all over the news: “AI will take over all our jobs”, “mass layoffs because of AI”, “anxiety whether humans will be…

    1 条评论
  • Imagination is the limit with ControlNet 1.1

    Imagination is the limit with ControlNet 1.1

    I’ve been backlogged on AI articles since so many new products come out every day, so it is probably time for me to…

  • Using LLaMa to Impersonate Friends

    Using LLaMa to Impersonate Friends

    (If you want to follow my AI learnings, you can follow my substack here) What if you could take a chat room with…

    1 条评论
  • AutoGPT is the next iteration of ChatGPT

    AutoGPT is the next iteration of ChatGPT

    Alright, we’ve heard lots of amazing uses for ChatGPT and how it will enhance (or even replace) human tasks. With the…

    1 条评论
  • The AI (3D) Space Race

    The AI (3D) Space Race

    Every day, the Generative AI and Art space has another breakthrough, whether it’s creating photorealistic images and…

  • The Open-Source LLM Effect

    The Open-Source LLM Effect

    Keeping up with AI is a full time job, I would know since I’m looking for one! Everyday, new start-ups are created…

    5 条评论
  • Thanks to LLaMa, you too can GPT

    Thanks to LLaMa, you too can GPT

    AI is moving at a blistering fast pace, there are mind-blowing Stable Diffusion extensions that come out every week…

  • First there was COIN, then there was BASE

    First there was COIN, then there was BASE

    (Excerpt taken from my newsletter at https://www.rocketfuelcrypto.

    1 条评论

社区洞察

其他会员也浏览了