Great post! o1/pro is the first model i've used that can do high level software architecture well:
- As you noted, give it all the context -- all relevant code files + existing design docs (RepoPrompt is great for this).
- Ramble about the problem into speech2text for a while
- At the end tell it to present multiple alternatives + reasons to use/not use
Breakthrough capability is lack of sycophancy -- it's the first model I've used where I disagree with it, and then it will hold its ground and convince me that it is right.
Another tip is to have it break up the implementation into discrete steps, outputting all context for each one. Then paste into cursor composer one at a time for the actual coding.
Someone else was saying that after each step, they go back to o1 and have it review the code that cursor wrote. Still need to try that one out!
How much disconnected rambling can o1 handle? Can I just speak stream of thought about all the discussions and back and forth ideas that happened for a product feature and dump them? I'm clear on the final output that I want. I'm just thinking of o1 can handle all this extra discussion context.
I was wondering if you could spare a moment and help me out. I am looking for a portable (or not) local LLM / SLM installation so I can build a LLM (or SLM - small language model) AI assistant (agent) on my laptop. I'll then feed it with various ebooks/data/articles and see how it can speed up (or improve) my learning?
Can you please suggest a solution if you know of any? I appreciate it very much!
I'm searching for a solution and right now reading through the below:
for local inference, many folks now use ollama. but you seem to also want a RAG UI - in which case the options you list probably work. personally, i dont care about local, and use Claude Projects or NotebookLM.
Great post! o1/pro is the first model i've used that can do high level software architecture well:
- As you noted, give it all the context -- all relevant code files + existing design docs (RepoPrompt is great for this).
- Ramble about the problem into speech2text for a while
- At the end tell it to present multiple alternatives + reasons to use/not use
Breakthrough capability is lack of sycophancy -- it's the first model I've used where I disagree with it, and then it will hold its ground and convince me that it is right.
Another tip is to have it break up the implementation into discrete steps, outputting all context for each one. Then paste into cursor composer one at a time for the actual coding.
Someone else was saying that after each step, they go back to o1 and have it review the code that cursor wrote. Still need to try that one out!
great responses, thanks for sharing
How does Gemini deep research compare?
Hmm you mention adding pdfs to the prompt, I don’t have this feature in GPT Plus (and not only me: https://www.reddit.com/r/OpenAI/comments/1hwli30/pdf_file_uploads_on_o1/). Is that a Pro feature?
How much disconnected rambling can o1 handle? Can I just speak stream of thought about all the discussions and back and forth ideas that happened for a product feature and dump them? I'm clear on the final output that I want. I'm just thinking of o1 can handle all this extra discussion context.
lots! i ramble all the time and just dump the transcript
o1 only accepts ~32k tokens right? So maybe dumping a bunch of thoughts might be better in a separate chat?
Source: llm-stats.com
Hi there.
Thanks for this article. :)
I was wondering if you could spare a moment and help me out. I am looking for a portable (or not) local LLM / SLM installation so I can build a LLM (or SLM - small language model) AI assistant (agent) on my laptop. I'll then feed it with various ebooks/data/articles and see how it can speed up (or improve) my learning?
Can you please suggest a solution if you know of any? I appreciate it very much!
I'm searching for a solution and right now reading through the below:
https://docs.anythingllm.com/installation-desktop/overview
https://medium.com/thedeephub/50-open-source-options-for-running-llms-locally-db1ec6f5a54f
https://semaphoreci.com/blog/local-llm
for local inference, many folks now use ollama. but you seem to also want a RAG UI - in which case the options you list probably work. personally, i dont care about local, and use Claude Projects or NotebookLM.
Thank you.
love it! thanks