11 Comments

Great post! o1/pro is the first model i've used that can do high level software architecture well:

- As you noted, give it all the context -- all relevant code files + existing design docs (RepoPrompt is great for this).

- Ramble about the problem into speech2text for a while

- At the end tell it to present multiple alternatives + reasons to use/not use

Breakthrough capability is lack of sycophancy -- it's the first model I've used where I disagree with it, and then it will hold its ground and convince me that it is right.

Another tip is to have it break up the implementation into discrete steps, outputting all context for each one. Then paste into cursor composer one at a time for the actual coding.

Someone else was saying that after each step, they go back to o1 and have it review the code that cursor wrote. Still need to try that one out!

Expand full comment

great responses, thanks for sharing

Expand full comment

How does Gemini deep research compare?

Expand full comment

Hmm you mention adding pdfs to the prompt, I don’t have this feature in GPT Plus (and not only me: https://www.reddit.com/r/OpenAI/comments/1hwli30/pdf_file_uploads_on_o1/). Is that a Pro feature?

Expand full comment

How much disconnected rambling can o1 handle? Can I just speak stream of thought about all the discussions and back and forth ideas that happened for a product feature and dump them? I'm clear on the final output that I want. I'm just thinking of o1 can handle all this extra discussion context.

Expand full comment

lots! i ramble all the time and just dump the transcript

Expand full comment

o1 only accepts ~32k tokens right? So maybe dumping a bunch of thoughts might be better in a separate chat?

Source: llm-stats.com

Expand full comment

Hi there.

Thanks for this article. :)

I was wondering if you could spare a moment and help me out. I am looking for a portable (or not) local LLM / SLM installation so I can build a LLM (or SLM - small language model) AI assistant (agent) on my laptop. I'll then feed it with various ebooks/data/articles and see how it can speed up (or improve) my learning?

Can you please suggest a solution if you know of any? I appreciate it very much!

I'm searching for a solution and right now reading through the below:

https://docs.anythingllm.com/installation-desktop/overview

https://medium.com/thedeephub/50-open-source-options-for-running-llms-locally-db1ec6f5a54f

https://semaphoreci.com/blog/local-llm

Expand full comment

for local inference, many folks now use ollama. but you seem to also want a RAG UI - in which case the options you list probably work. personally, i dont care about local, and use Claude Projects or NotebookLM.

Expand full comment

Thank you.

Expand full comment

love it! thanks

Expand full comment