fbpx
Wikipedia

Syntax (programming languages)

In computer science, the syntax of a computer language is the rules that define the combinations of symbols that are considered to be correctly structured statements or expressions in that language. This applies both to programming languages, where the document represents source code, and to markup languages, where the document represents data.

Syntax highlighting and indent style are often used to aid programmers in recognizing elements of source code. This Python code uses color coded highlighting.

The syntax of a language defines its surface form.[1] Text-based computer languages are based on sequences of characters, while visual programming languages are based on the spatial layout and connections between symbols (which may be textual or graphical). Documents that are syntactically invalid are said to have a syntax error. When designing the syntax of a language, a designer might start by writing down examples of both legal and illegal strings, before trying to figure out the general rules from these examples.[2]

Syntax therefore refers to the form of the code, and is contrasted with semantics – the meaning. In processing computer languages, semantic processing generally comes after syntactic processing; however, in some cases, semantic processing is necessary for complete syntactic analysis, and these are done together or concurrently. In a compiler, the syntactic analysis comprises the frontend, while the semantic analysis comprises the backend (and middle end, if this phase is distinguished).

Levels of syntax edit

Computer language syntax is generally distinguished into three levels:

  • Words – the lexical level, determining how characters form tokens;
  • Phrases – the grammar level, narrowly speaking, determining how tokens form phrases;
  • Context – determining what objects or variables names refer to, if types are valid, etc.

Distinguishing in this way yields modularity, allowing each level to be described and processed separately and often independently.

First, a lexer turns the linear sequence of characters into a linear sequence of tokens; this is known as "lexical analysis" or "lexing".[3]

Second, the parser turns the linear sequence of tokens into a hierarchical syntax tree; this is known as "parsing" narrowly speaking. This ensures that the line of tokens conform to the formal grammars of the programming language. The parsing stage itself can be divided into two parts: the parse tree, or "concrete syntax tree", which is determined by the grammar, but is generally far too detailed for practical use, and the abstract syntax tree (AST), which simplifies this into a usable form. The AST and contextual analysis steps can be considered a form of semantic analysis, as they are adding meaning and interpretation to the syntax, or alternatively as informal, manual implementations of syntactical rules that would be difficult or awkward to describe or implement formally.

Thirdly, the contextual analysis resolves names and checks types. This modularity is sometimes possible, but in many real-world languages an earlier step depends on a later step – for example, the lexer hack in C is because tokenization depends on context. Even in these cases, syntactical analysis is often seen as approximating this ideal model.

The levels generally correspond to levels in the Chomsky hierarchy. Words are in a regular language, specified in the lexical grammar, which is a Type-3 grammar, generally given as regular expressions. Phrases are in a context-free language (CFL), generally a deterministic context-free language (DCFL), specified in a phrase structure grammar, which is a Type-2 grammar, generally given as production rules in Backus–Naur form (BNF). Phrase grammars are often specified in much more constrained grammars than full context-free grammars, in order to make them easier to parse; while the LR parser can parse any DCFL in linear time, the simple LALR parser and even simpler LL parser are more efficient, but can only parse grammars whose production rules are constrained. In principle, contextual structure can be described by a context-sensitive grammar, and automatically analyzed by means such as attribute grammars, though, in general, this step is done manually, via name resolution rules and type checking, and implemented via a symbol table which stores names and types for each scope.

Tools have been written that automatically generate a lexer from a lexical specification written in regular expressions and a parser from the phrase grammar written in BNF: this allows one to use declarative programming, rather than need to have procedural or functional programming. A notable example is the lex-yacc pair. These automatically produce a concrete syntax tree; the parser writer must then manually write code describing how this is converted to an abstract syntax tree. Contextual analysis is also generally implemented manually. Despite the existence of these automatic tools, parsing is often implemented manually, for various reasons – perhaps the phrase structure is not context-free, or an alternative implementation improves performance or error-reporting, or allows the grammar to be changed more easily. Parsers are often written in functional languages, such as Haskell, or in scripting languages, such as Python or Perl, or in C or C++.

Examples of errors edit

As an example, (add 1 1) is a syntactically valid Lisp program (assuming the 'add' function exists, else name resolution fails), adding 1 and 1. However, the following are invalid:

(_ 1 1) lexical error: '_' is not valid (add 1 1 parsing error: missing closing ')' 

The lexer is unable to identify the first error – all it knows is that, after producing the token LEFT_PAREN, '(' the remainder of the program is invalid, since no word rule begins with '_'. The second error is detected at the parsing stage: The parser has identified the "list" production rule due to the '(' token (as the only match), and thus can give an error message; in general it may be ambiguous.

Type errors and undeclared variable errors are sometimes considered to be syntax errors when they are detected at compile-time (which is usually the case when compiling strongly-typed languages), though it is common to classify these kinds of error as semantic errors instead.[4][5][6]

As an example, the Python code

'a' + 1 

contains a type error because it adds a string literal to an integer literal. Type errors of this kind can be detected at compile-time: They can be detected during parsing (phrase analysis) if the compiler uses separate rules that allow "integerLiteral + integerLiteral" but not "stringLiteral + integerLiteral", though it is more likely that the compiler will use a parsing rule that allows all expressions of the form "LiteralOrIdentifier + LiteralOrIdentifier" and then the error will be detected during contextual analysis (when type checking occurs). In some cases this validation is not done by the compiler, and these errors are only detected at runtime.

In a dynamically typed language, where type can only be determined at runtime, many type errors can only be detected at runtime. For example, the Python code

a + b 

is syntactically valid at the phrase level, but the correctness of the types of a and b can only be determined at runtime, as variables do not have types in Python, only values do. Whereas there is disagreement about whether a type error detected by the compiler should be called a syntax error (rather than a static semantic error), type errors which can only be detected at program execution time are always regarded as semantic rather than syntax errors.

Syntax definition edit

 
Parse tree of Python code with inset tokenization

The syntax of textual programming languages is usually defined using a combination of regular expressions (for lexical structure) and Backus–Naur form (a metalanguage for grammatical structure) to inductively specify syntactic categories (nonterminal) and terminal symbols.[7] Syntactic categories are defined by rules called productions, which specify the values that belong to a particular syntactic category.[1] Terminal symbols are the concrete characters or strings of characters (for example keywords such as define, if, let, or void) from which syntactically valid programs are constructed.

Syntax can be divided into context-free syntax and context-sensitive syntax.[7] Context-free syntax are rules directed by the metalanguage of the programming language. These would not be constrained by the context surrounding or referring that part of the syntax, whereas context-sensitive syntax would.

A language can have different equivalent grammars, such as equivalent regular expressions (at the lexical levels), or different phrase rules which generate the same language. Using a broader category of grammars, such as LR grammars, can allow shorter or simpler grammars compared with more restricted categories, such as LL grammar, which may require longer grammars with more rules. Different but equivalent phrase grammars yield different parse trees, though the underlying language (set of valid documents) is the same.

Example: Lisp S-expressions edit

Below is a simple grammar, defined using the notation of regular expressions and Extended Backus–Naur form. It describes the syntax of S-expressions, a data syntax of the programming language Lisp, which defines productions for the syntactic categories expression, atom, number, symbol, and list:

expression = atom | list atom = number | symbol  number = [+-]?['0'-'9']+ symbol = ['A'-'Z']['A'-'Z''0'-'9'].* list = '(', expression*, ')' 

This grammar specifies the following:

  • an expression is either an atom or a list;
  • an atom is either a number or a symbol;
  • a number is an unbroken sequence of one or more decimal digits, optionally preceded by a plus or minus sign;
  • a symbol is a letter followed by zero or more of any characters (excluding whitespace); and
  • a list is a matched pair of parentheses, with zero or more expressions inside it.

Here the decimal digits, upper- and lower-case characters, and parentheses are terminal symbols.

The following are examples of well-formed token sequences in this grammar: '12345', '()', '(A B C232 (1))'

Complex grammars edit

The grammar needed to specify a programming language can be classified by its position in the Chomsky hierarchy. The phrase grammar of most programming languages can be specified using a Type-2 grammar, i.e., they are context-free grammars,[8] though the overall syntax is context-sensitive (due to variable declarations and nested scopes), hence Type-1. However, there are exceptions, and for some languages the phrase grammar is Type-0 (Turing-complete).

In some languages like Perl and Lisp the specification (or implementation) of the language allows constructs that execute during the parsing phase. Furthermore, these languages have constructs that allow the programmer to alter the behavior of the parser. This combination effectively blurs the distinction between parsing and execution, and makes syntax analysis an undecidable problem in these languages, meaning that the parsing phase may not finish. For example, in Perl it is possible to execute code during parsing using a BEGIN statement, and Perl function prototypes may alter the syntactic interpretation, and possibly even the syntactic validity of the remaining code.[9][10] Colloquially this is referred to as "only Perl can parse Perl" (because code must be executed during parsing, and can modify the grammar), or more strongly "even Perl cannot parse Perl" (because it is undecidable). Similarly, Lisp macros introduced by the defmacro syntax also execute during parsing, meaning that a Lisp compiler must have an entire Lisp run-time system present. In contrast, C macros are merely string replacements, and do not require code execution.[11][12]

Syntax versus semantics edit

The syntax of a language describes the form of a valid program, but does not provide any information about the meaning of the program or the results of executing that program. The meaning given to a combination of symbols is handled by semantics (either formal or hard-coded in a reference implementation). Valid syntax must be established before semantics can make meaning out of it.[7] Not all syntactically correct programs are semantically correct. Many syntactically correct programs are nonetheless ill-formed, per the language's rules; and may (depending on the language specification and the soundness of the implementation) result in an error on translation or execution. In some cases, such programs may exhibit undefined behavior. Even when a program is well-defined within a language, it may still have a meaning that is not intended by the person who wrote it.

Using natural language as an example, it may not be possible to assign a meaning to a grammatically correct sentence or the sentence may be false:

  • "Colorless green ideas sleep furiously." is grammatically well formed but has no generally accepted meaning.
  • "John is a married bachelor." is grammatically well formed but expresses a meaning that cannot be true.

The following C language fragment is syntactically correct, but performs an operation that is not semantically defined (because p is a null pointer, the operations p->real and p->im have no meaning):

 complex *p = NULL;  complex abs_p = sqrt (p->real * p->real + p->im * p->im); 

As a simpler example,

 int x;  printf("%d", x); 

is syntactically valid, but not semantically defined, as it uses an uninitialized variable. Even though compilers for some programming languages (e.g., Java and C#) would detect uninitialized variable errors of this kind, they should be regarded as semantic errors rather than syntax errors.[6][13]

See also edit

To quickly compare syntax of various programming languages, take a look at the list of "Hello, World!" program examples:

References edit

  1. ^ a b Friedman, Daniel P.; Mitchell Wand; Christopher T. Haynes (1992). Essentials of Programming Languages (1st ed.). The MIT Press. ISBN 0-262-06145-7.
  2. ^ Smith, Dennis (1999). Designing Maintainable Software. Springer Science & Business Media.
  3. ^ Pai, Vaikunta; Aithal, P.S. (December 31, 2020). "A Systematic Literature Review of Lexical Analyzer Implementation Techniques in Compiler Design". International Journal of Applied Engineering and Management Letters (IJAEML). 4 (2): 285–301. ISSN 2581-7000.
  4. ^ Aho, Alfred V.; Monica S. Lam; Ravi Sethi; Jeffrey D. Ullman (2007). Compilers: Principles, Techniques, and Tools (2nd ed.). Addison Wesley. ISBN 0-321-48681-1.Section 4.1.3: Syntax Error Handling, pp.194–195.
  5. ^ Louden, Kenneth C. (1997). Compiler Construction: Principles and Practice. Brooks/Cole. ISBN 981-243-694-4. Exercise 1.3, pp.27–28.
  6. ^ a b Semantic Errors in Java
  7. ^ a b c Sloneggger, Kenneth; Kurtz, Barry (1995). Formal Syntax and Semantics of Programming Languages. Addison-Wesley Publishing Company. ISBN 0-201-65697-3.{{cite book}}: CS1 maint: date and year (link)
  8. ^ Michael Sipser (1997). "2.2 Pushdown Automata". Introduction to the Theory of Computation. PWS Publishing. pp. 101–114. ISBN 0-534-94728-X.
  9. ^ LtU comment clarifying that the undecidable problem is membership in the class of Perl programs
  10. ^ chromatic's example of Perl code that gives a syntax error depending on the value of random variable
  11. ^ . Apl.jhu.edu. 1996-02-08. Archived from the original on 2013-08-06. Retrieved 2013-08-17.
  12. ^ "The Common Lisp Cookbook - Macros and Backquote". Cl-cookbook.sourceforge.net. 2007-01-16. Retrieved 2013-08-17.
  13. ^ Issue of syntax or semantics?

External links edit

  • Various syntactic constructs used in computer programming languages
  • Python error “ImportError: No module named” Why? How? Command-Line? [Solved2021]

syntax, programming, languages, this, article, needs, additional, citations, verification, please, help, improve, this, article, adding, citations, reliable, sources, unsourced, material, challenged, removed, find, sources, syntax, programming, languages, news. This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Syntax programming languages news newspapers books scholar JSTOR August 2013 Learn how and when to remove this template message In computer science the syntax of a computer language is the rules that define the combinations of symbols that are considered to be correctly structured statements or expressions in that language This applies both to programming languages where the document represents source code and to markup languages where the document represents data Syntax highlighting and indent style are often used to aid programmers in recognizing elements of source code This Python code uses color coded highlighting The syntax of a language defines its surface form 1 Text based computer languages are based on sequences of characters while visual programming languages are based on the spatial layout and connections between symbols which may be textual or graphical Documents that are syntactically invalid are said to have a syntax error When designing the syntax of a language a designer might start by writing down examples of both legal and illegal strings before trying to figure out the general rules from these examples 2 Syntax therefore refers to the form of the code and is contrasted with semantics the meaning In processing computer languages semantic processing generally comes after syntactic processing however in some cases semantic processing is necessary for complete syntactic analysis and these are done together or concurrently In a compiler the syntactic analysis comprises the frontend while the semantic analysis comprises the backend and middle end if this phase is distinguished Contents 1 Levels of syntax 1 1 Examples of errors 2 Syntax definition 2 1 Example Lisp S expressions 2 2 Complex grammars 3 Syntax versus semantics 4 See also 5 References 6 External linksLevels of syntax editComputer language syntax is generally distinguished into three levels Words the lexical level determining how characters form tokens Phrases the grammar level narrowly speaking determining how tokens form phrases Context determining what objects or variables names refer to if types are valid etc Distinguishing in this way yields modularity allowing each level to be described and processed separately and often independently First a lexer turns the linear sequence of characters into a linear sequence of tokens this is known as lexical analysis or lexing 3 Second the parser turns the linear sequence of tokens into a hierarchical syntax tree this is known as parsing narrowly speaking This ensures that the line of tokens conform to the formal grammars of the programming language The parsing stage itself can be divided into two parts the parse tree or concrete syntax tree which is determined by the grammar but is generally far too detailed for practical use and the abstract syntax tree AST which simplifies this into a usable form The AST and contextual analysis steps can be considered a form of semantic analysis as they are adding meaning and interpretation to the syntax or alternatively as informal manual implementations of syntactical rules that would be difficult or awkward to describe or implement formally Thirdly the contextual analysis resolves names and checks types This modularity is sometimes possible but in many real world languages an earlier step depends on a later step for example the lexer hack in C is because tokenization depends on context Even in these cases syntactical analysis is often seen as approximating this ideal model The levels generally correspond to levels in the Chomsky hierarchy Words are in a regular language specified in the lexical grammar which is a Type 3 grammar generally given as regular expressions Phrases are in a context free language CFL generally a deterministic context free language DCFL specified in a phrase structure grammar which is a Type 2 grammar generally given as production rules in Backus Naur form BNF Phrase grammars are often specified in much more constrained grammars than full context free grammars in order to make them easier to parse while the LR parser can parse any DCFL in linear time the simple LALR parser and even simpler LL parser are more efficient but can only parse grammars whose production rules are constrained In principle contextual structure can be described by a context sensitive grammar and automatically analyzed by means such as attribute grammars though in general this step is done manually via name resolution rules and type checking and implemented via a symbol table which stores names and types for each scope Tools have been written that automatically generate a lexer from a lexical specification written in regular expressions and a parser from the phrase grammar written in BNF this allows one to use declarative programming rather than need to have procedural or functional programming A notable example is the lex yacc pair These automatically produce a concrete syntax tree the parser writer must then manually write code describing how this is converted to an abstract syntax tree Contextual analysis is also generally implemented manually Despite the existence of these automatic tools parsing is often implemented manually for various reasons perhaps the phrase structure is not context free or an alternative implementation improves performance or error reporting or allows the grammar to be changed more easily Parsers are often written in functional languages such as Haskell or in scripting languages such as Python or Perl or in C or C Examples of errors edit Main article Syntax error As an example add 1 1 is a syntactically valid Lisp program assuming the add function exists else name resolution fails adding 1 and 1 However the following are invalid 1 1 lexical error is not valid add 1 1 parsing error missing closing The lexer is unable to identify the first error all it knows is that after producing the token LEFT PAREN the remainder of the program is invalid since no word rule begins with The second error is detected at the parsing stage The parser has identified the list production rule due to the token as the only match and thus can give an error message in general it may be ambiguous Type errors and undeclared variable errors are sometimes considered to be syntax errors when they are detected at compile time which is usually the case when compiling strongly typed languages though it is common to classify these kinds of error as semantic errors instead 4 5 6 As an example the Python code a 1 contains a type error because it adds a string literal to an integer literal Type errors of this kind can be detected at compile time They can be detected during parsing phrase analysis if the compiler uses separate rules that allow integerLiteral integerLiteral but not stringLiteral integerLiteral though it is more likely that the compiler will use a parsing rule that allows all expressions of the form LiteralOrIdentifier LiteralOrIdentifier and then the error will be detected during contextual analysis when type checking occurs In some cases this validation is not done by the compiler and these errors are only detected at runtime In a dynamically typed language where type can only be determined at runtime many type errors can only be detected at runtime For example the Python code a b is syntactically valid at the phrase level but the correctness of the types of a and b can only be determined at runtime as variables do not have types in Python only values do Whereas there is disagreement about whether a type error detected by the compiler should be called a syntax error rather than a static semantic error type errors which can only be detected at program execution time are always regarded as semantic rather than syntax errors Syntax definition edit nbsp Parse tree of Python code with inset tokenizationThe syntax of textual programming languages is usually defined using a combination of regular expressions for lexical structure and Backus Naur form a metalanguage for grammatical structure to inductively specify syntactic categories nonterminal and terminal symbols 7 Syntactic categories are defined by rules called productions which specify the values that belong to a particular syntactic category 1 Terminal symbols are the concrete characters or strings of characters for example keywords such as define if let or void from which syntactically valid programs are constructed Syntax can be divided into context free syntax and context sensitive syntax 7 Context free syntax are rules directed by the metalanguage of the programming language These would not be constrained by the context surrounding or referring that part of the syntax whereas context sensitive syntax would A language can have different equivalent grammars such as equivalent regular expressions at the lexical levels or different phrase rules which generate the same language Using a broader category of grammars such as LR grammars can allow shorter or simpler grammars compared with more restricted categories such as LL grammar which may require longer grammars with more rules Different but equivalent phrase grammars yield different parse trees though the underlying language set of valid documents is the same Example Lisp S expressions edit Below is a simple grammar defined using the notation of regular expressions and Extended Backus Naur form It describes the syntax of S expressions a data syntax of the programming language Lisp which defines productions for the syntactic categories expression atom number symbol and list expression atom list atom number symbol number 0 9 symbol A Z A Z 0 9 list expression This grammar specifies the following an expression is either an atom or a list an atom is either a number or a symbol a number is an unbroken sequence of one or more decimal digits optionally preceded by a plus or minus sign a symbol is a letter followed by zero or more of any characters excluding whitespace and a list is a matched pair of parentheses with zero or more expressions inside it Here the decimal digits upper and lower case characters and parentheses are terminal symbols The following are examples of well formed token sequences in this grammar 12345 A B C232 1 Complex grammars edit The grammar needed to specify a programming language can be classified by its position in the Chomsky hierarchy The phrase grammar of most programming languages can be specified using a Type 2 grammar i e they are context free grammars 8 though the overall syntax is context sensitive due to variable declarations and nested scopes hence Type 1 However there are exceptions and for some languages the phrase grammar is Type 0 Turing complete In some languages like Perl and Lisp the specification or implementation of the language allows constructs that execute during the parsing phase Furthermore these languages have constructs that allow the programmer to alter the behavior of the parser This combination effectively blurs the distinction between parsing and execution and makes syntax analysis an undecidable problem in these languages meaning that the parsing phase may not finish For example in Perl it is possible to execute code during parsing using a BEGIN statement and Perl function prototypes may alter the syntactic interpretation and possibly even the syntactic validity of the remaining code 9 10 Colloquially this is referred to as only Perl can parse Perl because code must be executed during parsing and can modify the grammar or more strongly even Perl cannot parse Perl because it is undecidable Similarly Lisp macros introduced by the defmacro syntax also execute during parsing meaning that a Lisp compiler must have an entire Lisp run time system present In contrast C macros are merely string replacements and do not require code execution 11 12 Syntax versus semantics editThe syntax of a language describes the form of a valid program but does not provide any information about the meaning of the program or the results of executing that program The meaning given to a combination of symbols is handled by semantics either formal or hard coded in a reference implementation Valid syntax must be established before semantics can make meaning out of it 7 Not all syntactically correct programs are semantically correct Many syntactically correct programs are nonetheless ill formed per the language s rules and may depending on the language specification and the soundness of the implementation result in an error on translation or execution In some cases such programs may exhibit undefined behavior Even when a program is well defined within a language it may still have a meaning that is not intended by the person who wrote it Using natural language as an example it may not be possible to assign a meaning to a grammatically correct sentence or the sentence may be false Colorless green ideas sleep furiously is grammatically well formed but has no generally accepted meaning John is a married bachelor is grammatically well formed but expresses a meaning that cannot be true The following C language fragment is syntactically correct but performs an operation that is not semantically defined because span class n p span is a null pointer the operations span class n p span span class o gt span span class n real span and span class n p span span class o gt span span class n im span have no meaning complex p NULL complex abs p sqrt p gt real p gt real p gt im p gt im As a simpler example int x printf d x is syntactically valid but not semantically defined as it uses an uninitialized variable Even though compilers for some programming languages e g Java and C would detect uninitialized variable errors of this kind they should be regarded as semantic errors rather than syntax errors 6 13 See also editNaming convention programming Comparison of programming languages syntax To quickly compare syntax of various programming languages take a look at the list of Hello World program examples Prolog syntax and semantics Perl syntax PHP syntax and semantics C syntax C syntax Java syntax JavaScript syntax Python syntax and semantics Lua syntax Haskell syntaxReferences edit a b Friedman Daniel P Mitchell Wand Christopher T Haynes 1992 Essentials of Programming Languages 1st ed The MIT Press ISBN 0 262 06145 7 Smith Dennis 1999 Designing Maintainable Software Springer Science amp Business Media Pai Vaikunta Aithal P S December 31 2020 A Systematic Literature Review of Lexical Analyzer Implementation Techniques in Compiler Design International Journal of Applied Engineering and Management Letters IJAEML 4 2 285 301 ISSN 2581 7000 Aho Alfred V Monica S Lam Ravi Sethi Jeffrey D Ullman 2007 Compilers Principles Techniques and Tools 2nd ed Addison Wesley ISBN 0 321 48681 1 Section 4 1 3 Syntax Error Handling pp 194 195 Louden Kenneth C 1997 Compiler Construction Principles and Practice Brooks Cole ISBN 981 243 694 4 Exercise 1 3 pp 27 28 a b Semantic Errors in Java a b c Sloneggger Kenneth Kurtz Barry 1995 Formal Syntax and Semantics of Programming Languages Addison Wesley Publishing Company ISBN 0 201 65697 3 a href Template Cite book html title Template Cite book cite book a CS1 maint date and year link Michael Sipser 1997 2 2 Pushdown Automata Introduction to the Theory of Computation PWS Publishing pp 101 114 ISBN 0 534 94728 X LtU comment clarifying that the undecidable problem is membership in the class of Perl programs chromatic s example of Perl code that gives a syntax error depending on the value of random variable An Introduction to Common Lisp Macros Apl jhu edu 1996 02 08 Archived from the original on 2013 08 06 Retrieved 2013 08 17 The Common Lisp Cookbook Macros and Backquote Cl cookbook sourceforge net 2007 01 16 Retrieved 2013 08 17 Issue of syntax or semantics External links editVarious syntactic constructs used in computer programming languages Python error ImportError No module named Why How Command Line Solved2021 Retrieved from https en wikipedia org w index php title Syntax programming languages amp oldid 1213425884, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.