Module_1_part2_NLP
Module_1_part2_NLP
phonet
ic Form (PF)
Move α Elements are moved to form grammatical sentences "Mary is liked by John."
This structure shows that "ate" is the main verb, "the food" is the object, and
"in a dhaba" is an optional prepositional phrase modifying the VP.
AP (Adjective Phrase) Structure
Tree Explanation:
The preposition "in" is the Head.
The NP "a dhaba" is the Complement.
"a" is the determiner (Det).
"dhaba" is the noun (N).
👉 This structure shows that "in" is the
preposition, which takes "a dhaba" as
its complement.
Maximal Projection of Sentence Structure
S̅ (S-Bar or CP -
Complementizer
Phrase)
COMP
(Complementizer)
It explains why "She ate food" is correct but "She slept the bed" is
wrong.
•"Slept on the bed" is correct because "on the bed" is a PP, not an NP.
Projection Principle
•If (+p) → Acts like a pronoun (can take the role of "he," "she," "it," etc.).
•Example: pro, PRO
ಓದುತ್ತಾನೆ
refers to a person (like "he/she"), it behaves like a pronoun → +p
•If (-p) → Does NOT act like a pronoun (doesn’t refer to a person/thing).
•Example: Traces (wh-trace, NP-trace)
The cake was eaten tₙₚ.
The trace tₙₚ is just a placeholder, NOT a pronoun → -p
Wh-trace (t𝑤ℎ) – Wh-Movement
The subject (ನಾನು / "I") is not spoken but understood. This missing subject is represented as
small pro in GB theory:
Big PRO (Pronoun)
•The term PRO (uppercase) refers to an empty subject in
control constructions.
•It is called "Big" because it appears in non-finite clauses
(clauses without tense), unlike small pro, which appears in
finite clauses.
•PRO is "big" because it has more syntactic restrictions
than small pro.
Example
The teacher told Vishnu [PRO to study].
The meaning is: The teacher told Vishnu that Vishnu should study
Incorrect: The teacher told Vishnu [he to study]
Binding Theory
• Binding defined as
α binds ß iff :
α C-commands ß
α and ß are Co-indexed.
Eg : Mukesh was killed
[e1 INFL kill Mukesh ]
[Mukesh was killed (by
ei)]
Mukesh was killed.
Empty clause (ei) and
mukesh (Npi) are bound.
Binding theory can be
given as follows:
(a) An anaphor (+a) is
bound in its
1. Principle A (Anaphors: +a, -p)
An anaphor (e.g., "himself", "herself") must be bound in its government
category.
• Example:
• Correct: John saw himself. → "Himself" is bound by "John" in the same clause.
• Incorrect: John said that Mary saw himself. → "Himself" has no local binder
(wrong).
2. Principle B (Pronominals: -a, +p)
A pronominal (e.g., "he", "she") must be free in its government category.
• Example:
• Correct: John said that he left. → "He" is free in the embedded clause.
• Incorrect: John saw him. → "Him" is bound in the same clause (wrong).
3. Principle C (R-expressions: -a, -p)
An R-expression (Referential) (e.g., "Mukesh", "John", "Mary") must be free (not
bound) everywhere.
• Example:
• Correct: John said that Mary left. → "Mary" is not bound.
• Incorrect: He said that John left. → If "He" refers to "John", it's wrong because
"John" must be free.
Empty category Principle (ECP):
α properly governs ß iff :
α governs ß and α is lexical ( i.e, N V A or P) or
Α locally A – binds ß ECP says ‘ A trace must be properly governed’
Example: What did John eat __?“
(Wrong) What do you think that John ate __?
• In GB case theory deals with the distributions of NPs and mentions that each NP
must assigned a case.
(Case refers to the grammatical role that a noun (or pronoun) plays in a sentence.
Ex: She ate an apple)
• In English we have nominative, objective, genitive etc., cases which are assigned to
NPs at particular positions.
• Indian languages are rich in case markers, which are carried even
during movements.
Case Filter :
An NP is un grammatical if it has phonetic content or if it is an argument and
is not case marked.
ATN (Argument Transition Networks )- Computational model linking syntax and meaning
which used phrase structure trees to represent the surface of sentences and underlying
predicate –argument structure.
•↑ (Up Arrow): Refers to the f-structure of the larger node (the phrase
containing the element).
•↓ (Down Arrow): Refers to the f-structure of the current node (the specific
element itself).
Rule 1 (S → NP VP):
•(↑ subj = ↓) → This means that the f-structure of NP (subject) goes into the
f-structure of the entire sentence (S).
•(↑ = ↓) → This means that the f-structure of VP (verb phrase) directly
assigned to the f-structure of S (since the verb defines the sentence's
action).
Functional Notation in Lexical Functional Grammar (LFG)
f-structure of given sentence
Passive Sentences (The object becomes the subject, and the original subject moves to
an optional phrase.)
Example:
• Passive: The food was eaten by Tara. → ("The food" is now the
subject, "was eaten" is the verb, and "by Tara" is an optional agent
phrase.)
Active Structure: Pred = ‘eat < (↑ Subj) (↑ Obj) >’
Passive Structure: Pred = ‘eat < (↑ Obl_ag) (↑ Subj) >’
[oblique agent phrase (Obl_ag): special grammatical element used in passive
sentences to indicate who performed the action. Example: Active Voice: Tara ate
the food. Passive Voice (Sentence Rewritten): The food was eaten by Tara. “by
Tara" is the oblique agent phrase (Obl_ag) because it represents the original subject
(Tara) and it is no longer the main subject of the sentence ]
Causativization (Making Someone Do Something)
LFG does not create empty categories like GB Theory. Instead, it uses functional structures
to maintain connections and Coordination.
The moved phrase (which picture) is still linked to its original position functionally.
Coordination refers to how different sentence elements are linked logically.
Example: Tara likes ‘tea and coffee’
"Which picture does Tara like__ most?“
1. Focus
• Represents the wh-word phrase, which is in focus.
• In this case, "Which picture" is the focus.
2. Pred (‘picture ⟨(Oblth)⟩’)
• The predicate (Pred) represents the main word of the phrase.
• Here, "picture" is the noun, and Oblth (oblique thematic) indicates that the picture is
related to something else (like an owner or subject).
3. Oblth(Oblique Thematic Role)
• It refers to an oblique phrase that provides additional information, such as who the
picture is related to.
• Contains:
• pred ‘PRO’ → PRO stands for a pronoun-like element.
• Refl + → Suggests reflexive or reference to something already mentioned.
4. Subj (Subject)
• pred ‘Tara’ → Identifies "Tara" as the subject of the sentence.
5. Obj (Object)
• The object is left empty, as the wh-word ("which picture") has moved.
6. Pred (‘like ⟨(↑Subj) (↑Obj)⟩’)
• Represents the verb "like", which takes a subject and an object.
Paninian Framework
Its a linguistic model based on Paninian Grammar (PG), which was written by Panini in 500 BC in Sanskrit (originally titled
Astadhyayi). Though originally designed for Sanskrit, this framework can be applied to other Indian languages and some Asian
languages.
Key Features
1. SOV Structure: Unlike English (which follows Subject-Verb-Object [SVO] order), most Asian languages follow Subject-
Object-Verb [SOV] order.
1. Example:
1. English (SVO): Tara likes apples.
2. Hindi (SOV): तारा को सेब पसंद है। (Tara ko seb pasand hai)
2. Inflectional Richness:
1. Many Indian languages rely on inflectional changes to convey grammatical relationships (e.g., tense, case, gender),
instead of relying solely on word order.
2. Example (in Sanskrit):
1. रामः ग्रामं गच्छति (Rāmaḥ grāmaṁ gacchati) → Rama goes to the village.
2. रामेण ग्रामः गम्यते (Rāmeṇa grāmaḥ gamyate) → The village is gone to by Rama.
3. Here, "Rama" (रामः / रामेण) changes its form based on its grammatical role.
3. Syntactic and Semantic Cues:
1. The Paninian framework focuses on meaning-based analysis rather than just word order, making it useful for analyzing
complex Indian languages.
4. Ongoing Research:
1. The framework is still being explored for its application to Indian languages, as many complexities remain to be
explained.
Some Important features of Indian languages
Layered representation of PG
• General GB considers deep structure , Surface and LF, where LF nearer to Semantics
• Paninion grammar frame work is said to be syntactico- semantic
surface layer to deep semantics by passing to intermediate layers.
•Language as a multi-layered process. You start with spoken words (surface level),
add grammatical roles (vibhakti level), determine who is doing what (karaka level),
and finally understand the real meaning (semantic level).
•Paninian Grammar follows this approach to ensure sentences preserve meaning
even if word order changes.
•This is useful in languages like Hindi and Sanskrit, where word order is flexible,
but meaning remains clear.
• Vibhakti means inflection, but here it refers to
word (noun, verb,or other)groups based
either on case endings, post positions or
compound verbs, or main and auxiliary verbs
etc,.
• Instead of talking NP,VP,AP,PP or … word
groups are formed based on various kinds of
markers. These markers are language specific
but all indian languages can be represented at
Vibhakti Leve.l
• Karaka Level means Case in GB these are theta
criterion etc.,.
• PG has its own way of defining karaka relations, these
relations based on word groups participate in the
activity denoted by the verb group(syntactic &
semantic as well).
KARAKA THEORY
• Central theme of PG framework, relations are assigned
based on the roles played by various participates in the
main activity.
• Roles are reflected in the case markers and post
position markers.
• Case relations we can find in english langauge, richness
of the case endings found in indian languages .
• Karakas such as Karta (subject), karma(object),
Karna(instrument),sampradhana(beneficary),
Apandan(seperation) and Adikhran (locus).
Issues in Panininan Grammar(PG)
𝑠=(𝑤1,𝑤2,𝑤3,...,𝑤𝑛) Where: 𝑤1is the first word, 𝑤2 is the second word, and
Solution is smoothing
Smoothing is a technique used to fix the zero-probability problem by
adjusting how probabilities are assigned. It does this by giving a small amount
of probability to unseen n-grams so that nothing has zero probability.
One of the simplest smoothing techniques is Add-One Smoothing, also called
Laplace Smoothing.
What is Add-One Smoothing?
Add-One Smoothing (Laplace Smoothing) is a simple method where
we add 1 to all n-gram counts before normalizing probabilities.