Communities

Writing
Writing
Codidact Meta
Codidact Meta
The Great Outdoors
The Great Outdoors
Photography & Video
Photography & Video
Scientific Speculation
Scientific Speculation
Cooking
Cooking
Electrical Engineering
Electrical Engineering
Judaism
Judaism
Languages & Linguistics
Languages & Linguistics
Software Development
Software Development
Mathematics
Mathematics
Christianity
Christianity
Code Golf
Code Golf
Music
Music
Physics
Physics
Linux Systems
Linux Systems
Power Users
Power Users
Tabletop RPGs
Tabletop RPGs
Community Proposals
Community Proposals
tag:snake search within a tag
answers:0 unanswered questions
user:xxxx search by author id
score:0.5 posts with 0.5+ score
"snake oil" exact phrase
votes:4 posts with 4+ votes
created:<1w created < 1 week ago
post_type:xxxx type of post
Search help
Notifications
Mark all as read See all your notifications »
Q&A

Post History

60%
+1 −0
Q&A What underlying principle is at play for how objective or subjective a natural language instruction is?

It seems like what you are hinting at is the degree to which an instruction contains the context required to understand it, answer it, and evaluate the answer. Moreover, the question hints at an ob...

posted 9mo ago by Eric Isaac‭  ·  edited 9mo ago by Eric Isaac‭

Answer
#2: Post edited by user avatar Eric Isaac‭ · 2024-03-23T17:31:48Z (9 months ago)
Grammar correction
  • It seems like what you are hinting at is the degree to which an instruction contains the context required to understand it, answer it, and evaluate the answer. Moreover, the question hints at an objective instruction being one that is almost completely self-contained in these aspects.
  • In the example you gave (Capitalize all letter S characters in a sentence), the prompt contain all the information about the subject acted upon, and, presumably, the computer the program is running on has the concept of characters/letters explicitly encoded into its operating system, including the concept of capitalization.
  • In this sense, LLM prompts already represent a small subset of natural language instructions. If we were in the same room, and I pointed to an object and told you to hand it to me, the instruction could be considered objective if we can both see the object, but it isn’t self-contained in the way it would need to be for an LLM because of the lack of shared context (assuming the LLM can’t see).
  • All this to say, I think the information you are looking for may be _Contextual_ vs. _Semantic_ in nature.
  • It seems like what you are hinting at is the degree to which an instruction contains the context required to understand it, answer it, and evaluate the answer. Moreover, the question hints at an objective instruction being one that is almost completely self-contained in these aspects.
  • In the example you gave (Capitalize all letter S characters in a sentence), the prompt contains all the information about the subject acted upon, and, presumably, the computer the program is running on has the concept of characters/letters explicitly encoded into its operating system, including the concept of capitalization.
  • In this sense, LLM prompts already represent a small subset of natural language instructions. If we were in the same room, and I pointed to an object and told you to hand it to me, the instruction could be considered objective if we can both see the object, but it isn’t self-contained in the way it would need to be for an LLM because of the lack of shared context (assuming the LLM can’t see).
  • All this to say, I think the information you are looking for may be _Contextual_ vs. _Semantic_ in nature.
#1: Initial revision by user avatar Eric Isaac‭ · 2024-03-23T12:25:36Z (9 months ago)
It seems like what you are hinting at is the degree to which an instruction contains the context required to understand it, answer it, and evaluate the answer. Moreover, the question hints at an objective instruction being one that is almost completely self-contained in these aspects.

In the example you gave (Capitalize all letter S characters in a sentence), the prompt contain all the information about the subject acted upon, and, presumably, the computer the program is running on has the concept of characters/letters explicitly encoded into its operating system, including the concept of capitalization.

In this sense, LLM prompts already represent a small subset of natural language instructions. If we were in the same room, and I pointed to an object and told you to hand it to me, the instruction could be considered objective if we can both see the object, but it isn’t self-contained in the way it would need to be for an LLM because of the lack of shared context (assuming the LLM can’t see).

All this to say, I think the information you are looking for may be _Contextual_ vs. _Semantic_ in nature.