The assignment for this project:
Use predictive models to generate text: either a Markov chain or a neural network, or both. How does your choice of source text affect the output? Try combining predictive text with other methods we’ve used for analyzing and generating text: use neural network-generated text to fill Tracery templates, or train a Markov model on the output of parsing parts of speech from a text, or some other combination. What works and what doesn’t? How does neural network-generated text “feel” different from Markov-generated text? How does the length of the n-gram and the unit of the n-gram affect the quality of the output?
This past weekend was the 2020 presidential election.
Each presents a different perspective and diagnosis of the election. A major topic that comes up in both is the loss of seats in the House for Democrats. They both make claims on reality - how things “really” are. That it’s just “common sense”. Subjective assertions that suddenly become concrete. Narratives seem to just emerge although someone has to drive the interviews.
Especially with news about politics, reading these articles can cause feeling a mix of dissonance and anxiety (Or is that just me then?).
I want to try and play with their language. The people interviewed and the person interviewing them. See if can recreate an interview article that generates similar feelings.
Note: the post continues inside the embed below
I also tried using aitextgen. I think training on the limited text I had didn’t work well. The text generated was way to close to the original. I also didn’t try a ton of messing with parameters because the markov chains were working well for me.
colab notebook (requires being logged into nyu google account)