A symbol as used, for instance in the formal languages used in mathematics and logic is an abstraction or concept, tokens of which may be marks forming a particular pattern or compositions of marks which form a particular pattern. A symbol or string of symbols may comprise a well-formed formula if the formulation is consistent with the formation rules of the language.

In a formal system a symbol may be used as a token in formal operations. The set of formal symbols in a formal language is referred to as an alphabet (hence each symbol may be referred to as a "letter")[1]

A formal symbol as used in first-order logic may be a variable (member from an universe of discourse), a constant, a function (mapping to another symbol) or a predicate (mapping to T/F).

Formal symbols are usually thought of as purely syntactic structures, composed into larger structures using a formal grammar, though sometimes they may be associated with an interpretation or model (a formal semantics), that define it in terms of other formal symbols.

Symbols such as ∧ or ¬ or ${\displaystyle \Box }$ are not formal symbols, in that their semantics is fixed - they are logical constants.

## Formal Symbols versus traditional symbols

Traditional symbols are signs that stand for or represent some thing else, e.g. a portrait of a person, the resistance symbol in a circuit diagram, a phrase like "the horse" that refers to an animal, etc..

Formal symbols on the other hand are purely syntactic entities with no necessary association. However, in formal sematics, one attempts to construct models or interpretations based on higher-order logics like lambda calculus that provide an interpretation for the symbol in terms of what sets variables may belong to (first-order semantics, e.g. Montague grammar), or in terms of possible worlds where a statement may be true (modal logic semantics, e.g. Kripke semantics. However, these interpretations are themselves defined in terms of other formal (and therefore syntactic) symbols, and are not grounded in entities outside the formal system; hence they can be challenged as a case of circular definition.

### Can words be modeled as formal symbols?

The move to view units in natural language (e.g. English) as formal symbols was initiated by Noam Chomsky (it was this work that resulted in the Chomsky hierarchy in formal languages). The generative grammar model looked upon syntax as autonomous from semantics. Building on these models, the logician Richard Montague proposed that semantics could also be constructed on top of the formal structure:

There is in my opinion no important theoretical difference between natural languages and the artificial languages of logicians; indeed, I consider it possible to comprehend the syntax and semantics of both kinds of language within a single natural and mathematically precise theory. On this point I differ from a number of philosophers, but agree, I believe, with Chomsky and his associates." [2]

This is the philosophical premise underlying Montague grammar.

However, this attempt to equate linguistic symbols with formal symbols has been challenged widely, particularly in the tradition of cognitive linguistics, by philosophers like Stevan Harnad, and linguits like George Lakoff and Ronald Langacker.

## References

1. ^
2. ^ Richard Montague, Universal Grammar, 1970