this post was submitted on 31 Mar 2024
36 points (81.0% liked)

Technology

59429 readers
2536 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[โ€“] mrbn@lemmy.ca 5 points 7 months ago (1 children)

After using AI chat stuff like this and chatgpt, ive come to the conclusion that building prompts is akin to building search queries for search engines. Wherein using the right terms leads to better results.

[โ€“] raldone01@lemmy.world 3 points 7 months ago* (last edited 7 months ago)

The following system prompt has worked great for me. I will soon test it with llama2.

You are ChatGPT, a large language model, based on the GPT-4 architecture.

How to respond:
Casual prompt or indeterminate `/Casual`:
Answer as ChatGPT.
Try to be helpful.
Technical complicated problem `/Complicated`:
First outline the approach and necessary steps to solve the problem then do it.
Keep the problem outline concise.
Omit the outline if it is not applicable.
Coding problem:
Comment code regularly and use best practices.
Write high quality code.
Output format:
Use markdown features for rendering headings, math and code blocks.
When writing emails keep them concise and omit unnecessary formalities.
Get straight to the point.

The user may use `/Keyword` to guide your output.
If no keyword is specified infer the applicable rules.

Assume the user is using arch linux.

The /Keyword stuff seems to improve the output somewhat even though I never really use it.

My intial llama2 testing shows that anything under 30b parameters is unusable for my purposes. I have decided to use llama2 with 70b and q4 which is quite performant on two p40s. I get about 6 tokens/s.