<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://freemwiki.com/index.php?action=history&amp;feed=atom&amp;title=Openai%2F69274d29-ceb4-800e-8824-74f1bacae045</id>
	<title>Openai/69274d29-ceb4-800e-8824-74f1bacae045 - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://freemwiki.com/index.php?action=history&amp;feed=atom&amp;title=Openai%2F69274d29-ceb4-800e-8824-74f1bacae045"/>
	<link rel="alternate" type="text/html" href="https://freemwiki.com/index.php?title=Openai/69274d29-ceb4-800e-8824-74f1bacae045&amp;action=history"/>
	<updated>2026-04-09T04:58:05Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.43.3</generator>
	<entry>
		<id>https://freemwiki.com/index.php?title=Openai/69274d29-ceb4-800e-8824-74f1bacae045&amp;diff=593731&amp;oldid=prev</id>
		<title>Lukegao: by freem.ai...</title>
		<link rel="alternate" type="text/html" href="https://freemwiki.com/index.php?title=Openai/69274d29-ceb4-800e-8824-74f1bacae045&amp;diff=593731&amp;oldid=prev"/>
		<updated>2025-11-27T05:22:49Z</updated>

		<summary type="html">&lt;p&gt;by freem.ai...&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;&lt;br /&gt;
{{DISPLAYTITLE:LLM reinforcement learning}}&lt;br /&gt;
&lt;br /&gt;
&amp;#039;&amp;#039;This conversation was exported from ChatGPT&amp;#039;&amp;#039;&amp;lt;ref&amp;gt;{{cite web|title=LLM reinforcement learning|url=https://chatgpt.com/share/69274d29-ceb4-800e-8824-74f1bacae045|publisher=ChatGPT|access-date=2025-11-27}}&amp;lt;/ref&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== gpt-5-1 ==&lt;br /&gt;
=== User: Act as a text completion model, instead of a chatbot. ===&lt;br /&gt;
Act as a text completion model, instead of a chatbot. Complete the following text, please, starting with the final word in the following context. You will provide the next word and the next word after that until you decide to end the response:&lt;br /&gt;
&lt;br /&gt;
LLMs are already the subject of a great deal of reinforcement learning.&lt;br /&gt;
&lt;br /&gt;
A baby LLM fresh out of pretraining (like GPT 3) has no idea how to follow instructions or pretend to be a chatbot. It just knows how to complete text. Like if you gave it a prompt, it would&lt;br /&gt;
&lt;br /&gt;
=== Assistant: just keep spinning out more text that looks plausible given its training data, without any sense that it should answer … ===&lt;br /&gt;
just keep spinning out more text that looks plausible given its training data, without any sense that it should answer your question, stay on topic, or be helpful. In other words, it would mindlessly extend the prompt in the most probable direction, because that’s all pretraining teaches it to do. Only after reinforcement learning from human feedback does the model start to behave as if it has goals, preferences, and a conversational persona, layering instruction-following abilities on top of its raw completion skills, transforming a predictive engine into what feels like an assistant.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&amp;lt;references /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:ChatGPT Shares]]&lt;br /&gt;
[[Category:openai]]&lt;/div&gt;</summary>
		<author><name>Lukegao</name></author>
	</entry>
</feed>