Communities

Writing
Writing
Codidact Meta
Codidact Meta
The Great Outdoors
The Great Outdoors
Photography & Video
Photography & Video
Scientific Speculation
Scientific Speculation
Cooking
Cooking
Electrical Engineering
Electrical Engineering
Judaism
Judaism
Languages & Linguistics
Languages & Linguistics
Software Development
Software Development
Mathematics
Mathematics
Christianity
Christianity
Code Golf
Code Golf
Music
Music
Physics
Physics
Linux Systems
Linux Systems
Power Users
Power Users
Tabletop RPGs
Tabletop RPGs
Community Proposals
Community Proposals
tag:snake search within a tag
answers:0 unanswered questions
user:xxxx search by author id
score:0.5 posts with 0.5+ score
"snake oil" exact phrase
votes:4 posts with 4+ votes
created:<1w created < 1 week ago
post_type:xxxx type of post
Search help
Notifications
Mark all as read See all your notifications »
Q&A

Comments on What underlying principle is at play for how objective or subjective a natural language instruction is?

Parent

What underlying principle is at play for how objective or subjective a natural language instruction is?

+2
−0

I am interested in exploring a series of prompts for a large language model which move from instructions which have a clear-cut "correct result", such as the instruction to capitalize every letter "S" in some sentence, to questions which may have a few acceptable results, to questions which are more open-ended and subjective.

I would like to think of some analytical framework which makes it clear exactly what is changing, presumably semantically, as we shift along that gradient. I can't see it clearly in my mind at the moment.

My guess is it could be modeled using information theory, and I can expand on how.

History
Why does this post require moderator attention?
You might want to add some details to your flag.
Why should this post be closed?

0 comment threads

Post
+1
−0

It seems like what you are hinting at is the degree to which an instruction contains the context required to understand it, answer it, and evaluate the answer. Moreover, the question hints at an objective instruction being one that is almost completely self-contained in these aspects.

In the example you gave (Capitalize all letter S characters in a sentence), the prompt contains all the information about the subject acted upon, and, presumably, the computer the program is running on has the concept of characters/letters explicitly encoded into its operating system, including the concept of capitalization.

In this sense, LLM prompts already represent a small subset of natural language instructions. If we were in the same room, and I pointed to an object and told you to hand it to me, the instruction could be considered objective if we can both see the object, but it isn’t self-contained in the way it would need to be for an LLM because of the lack of shared context (assuming the LLM can’t see).

All this to say, I think the information you are looking for may be Contextual vs. Semantic in nature.

History
Why does this post require moderator attention?
You might want to add some details to your flag.

1 comment thread

Misc. Thoughts (1 comment)
Misc. Thoughts
Eric Isaac‭ wrote about 2 months ago · edited about 2 months ago

I do want to say it seems like an interesting project! And it seems at a glance possible to model what you are hoping to model, I just don’t know how much luck you’ll have chasing down Semantics research to find the answers you are looking for.

It also seems like it’ll be important to communicate the restrictions on the set of instructions being considered (don’t present the theory as all encompassing for language instructions unless you are prepared to do an order of magnitude more research).